var/home/core/zuul-output/0000755000175000017500000000000015067167444014543 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067173401015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003635037115067173373017721 0ustar rootrootOct 01 09:09:30 crc systemd[1]: Starting Kubernetes Kubelet... Oct 01 09:09:30 crc restorecon[4667]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 09:09:30 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:09:31 crc restorecon[4667]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 01 09:09:32 crc kubenswrapper[4983]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.446375 4983 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455458 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455490 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455500 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455510 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455518 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455528 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455537 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455546 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455555 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455564 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455573 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455581 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455590 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455599 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455607 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455616 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455624 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455633 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455641 4983 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455650 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455659 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455667 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455675 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455684 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455693 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455701 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455720 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455730 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455738 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455750 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455762 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455771 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455783 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455795 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455838 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455851 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455860 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455869 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455877 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455886 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455898 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455908 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455918 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455927 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455936 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455945 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455957 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455966 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455975 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455983 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.455992 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456003 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456014 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456025 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456036 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456047 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456058 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456070 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456079 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456088 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456097 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456106 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456115 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456126 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456137 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456148 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456159 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456169 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456178 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456187 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.456197 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456381 4983 flags.go:64] FLAG: --address="0.0.0.0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456404 4983 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456427 4983 flags.go:64] FLAG: --anonymous-auth="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456443 4983 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456458 4983 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456468 4983 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456481 4983 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456496 4983 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456506 4983 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456516 4983 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456527 4983 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456537 4983 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456547 4983 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456557 4983 flags.go:64] FLAG: --cgroup-root="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456566 4983 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456576 4983 flags.go:64] FLAG: --client-ca-file="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456586 4983 flags.go:64] FLAG: --cloud-config="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456595 4983 flags.go:64] FLAG: --cloud-provider="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456605 4983 flags.go:64] FLAG: --cluster-dns="[]" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456618 4983 flags.go:64] FLAG: --cluster-domain="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456629 4983 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456640 4983 flags.go:64] FLAG: --config-dir="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456649 4983 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456660 4983 flags.go:64] FLAG: --container-log-max-files="5" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456683 4983 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456693 4983 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456703 4983 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456713 4983 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456723 4983 flags.go:64] FLAG: --contention-profiling="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456732 4983 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456742 4983 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456752 4983 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456762 4983 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456774 4983 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456784 4983 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456793 4983 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456834 4983 flags.go:64] FLAG: --enable-load-reader="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456845 4983 flags.go:64] FLAG: --enable-server="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456854 4983 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456871 4983 flags.go:64] FLAG: --event-burst="100" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456881 4983 flags.go:64] FLAG: --event-qps="50" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456890 4983 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456901 4983 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456911 4983 flags.go:64] FLAG: --eviction-hard="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456923 4983 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456933 4983 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456943 4983 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456953 4983 flags.go:64] FLAG: --eviction-soft="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456963 4983 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456973 4983 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456983 4983 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.456992 4983 flags.go:64] FLAG: --experimental-mounter-path="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457003 4983 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457012 4983 flags.go:64] FLAG: --fail-swap-on="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457021 4983 flags.go:64] FLAG: --feature-gates="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457033 4983 flags.go:64] FLAG: --file-check-frequency="20s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457042 4983 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457053 4983 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457062 4983 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457072 4983 flags.go:64] FLAG: --healthz-port="10248" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457082 4983 flags.go:64] FLAG: --help="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457092 4983 flags.go:64] FLAG: --hostname-override="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457101 4983 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457112 4983 flags.go:64] FLAG: --http-check-frequency="20s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457121 4983 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457131 4983 flags.go:64] FLAG: --image-credential-provider-config="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457140 4983 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457149 4983 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457159 4983 flags.go:64] FLAG: --image-service-endpoint="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457169 4983 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457178 4983 flags.go:64] FLAG: --kube-api-burst="100" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457189 4983 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457199 4983 flags.go:64] FLAG: --kube-api-qps="50" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457209 4983 flags.go:64] FLAG: --kube-reserved="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457218 4983 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457227 4983 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457237 4983 flags.go:64] FLAG: --kubelet-cgroups="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457247 4983 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457257 4983 flags.go:64] FLAG: --lock-file="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457267 4983 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457277 4983 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457287 4983 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457301 4983 flags.go:64] FLAG: --log-json-split-stream="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457311 4983 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457321 4983 flags.go:64] FLAG: --log-text-split-stream="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457331 4983 flags.go:64] FLAG: --logging-format="text" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457340 4983 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457350 4983 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457359 4983 flags.go:64] FLAG: --manifest-url="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457409 4983 flags.go:64] FLAG: --manifest-url-header="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457426 4983 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457439 4983 flags.go:64] FLAG: --max-open-files="1000000" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457454 4983 flags.go:64] FLAG: --max-pods="110" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457467 4983 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457480 4983 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457489 4983 flags.go:64] FLAG: --memory-manager-policy="None" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457499 4983 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457509 4983 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457518 4983 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457528 4983 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457548 4983 flags.go:64] FLAG: --node-status-max-images="50" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457558 4983 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457568 4983 flags.go:64] FLAG: --oom-score-adj="-999" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457580 4983 flags.go:64] FLAG: --pod-cidr="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457590 4983 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457605 4983 flags.go:64] FLAG: --pod-manifest-path="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457615 4983 flags.go:64] FLAG: --pod-max-pids="-1" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457624 4983 flags.go:64] FLAG: --pods-per-core="0" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457634 4983 flags.go:64] FLAG: --port="10250" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457644 4983 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457654 4983 flags.go:64] FLAG: --provider-id="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457663 4983 flags.go:64] FLAG: --qos-reserved="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457673 4983 flags.go:64] FLAG: --read-only-port="10255" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457683 4983 flags.go:64] FLAG: --register-node="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457693 4983 flags.go:64] FLAG: --register-schedulable="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457705 4983 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457721 4983 flags.go:64] FLAG: --registry-burst="10" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457731 4983 flags.go:64] FLAG: --registry-qps="5" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457740 4983 flags.go:64] FLAG: --reserved-cpus="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457749 4983 flags.go:64] FLAG: --reserved-memory="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457762 4983 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457772 4983 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457782 4983 flags.go:64] FLAG: --rotate-certificates="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457791 4983 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457800 4983 flags.go:64] FLAG: --runonce="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457839 4983 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457849 4983 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457859 4983 flags.go:64] FLAG: --seccomp-default="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457868 4983 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457878 4983 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457888 4983 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457898 4983 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457907 4983 flags.go:64] FLAG: --storage-driver-password="root" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457917 4983 flags.go:64] FLAG: --storage-driver-secure="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457926 4983 flags.go:64] FLAG: --storage-driver-table="stats" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457936 4983 flags.go:64] FLAG: --storage-driver-user="root" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457946 4983 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457956 4983 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457966 4983 flags.go:64] FLAG: --system-cgroups="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457975 4983 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457990 4983 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.457999 4983 flags.go:64] FLAG: --tls-cert-file="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458008 4983 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458021 4983 flags.go:64] FLAG: --tls-min-version="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458030 4983 flags.go:64] FLAG: --tls-private-key-file="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458040 4983 flags.go:64] FLAG: --topology-manager-policy="none" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458049 4983 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458059 4983 flags.go:64] FLAG: --topology-manager-scope="container" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458068 4983 flags.go:64] FLAG: --v="2" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458081 4983 flags.go:64] FLAG: --version="false" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458093 4983 flags.go:64] FLAG: --vmodule="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458112 4983 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.458122 4983 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458348 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458360 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458370 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458381 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458390 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458399 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458408 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458417 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458426 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458434 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458443 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458452 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458460 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458468 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458486 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458497 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458507 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458515 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458523 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458532 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458540 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458548 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458557 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458568 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458605 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458616 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458626 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458635 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458645 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458654 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458662 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458672 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458680 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458688 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458699 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458708 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458716 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458724 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458732 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458741 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458750 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458759 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458767 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458775 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458784 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458793 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458832 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458841 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458850 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458858 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458867 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458875 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458883 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458892 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458900 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458909 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458917 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458926 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458934 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458943 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458951 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458959 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458968 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458977 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458985 4983 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.458993 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.459005 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.459015 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.459026 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.459036 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.459047 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.459074 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.467842 4983 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.467885 4983 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.467978 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.467993 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468000 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468008 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468014 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468021 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468029 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468037 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468045 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468051 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468057 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468063 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468068 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468074 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468079 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468085 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468091 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468097 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468102 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468108 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468114 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468119 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468124 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468129 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468135 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468141 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468146 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468152 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468157 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468162 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468167 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468173 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468178 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468183 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468188 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468194 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468199 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468204 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468209 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468215 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468220 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468225 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468230 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468236 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468241 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468247 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468252 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468257 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468262 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468268 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468274 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468279 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468284 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468289 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468295 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468300 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468305 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468310 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468317 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468323 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468330 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468338 4983 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468343 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468348 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468354 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468360 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468366 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468372 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468377 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468382 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468387 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.468397 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468561 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468569 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468575 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468581 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468586 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468592 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468597 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468603 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468610 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468617 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468624 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468630 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468635 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468640 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468646 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468651 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468656 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468663 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468670 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468676 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468682 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468688 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468694 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468700 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468705 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468713 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468719 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468724 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468729 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468734 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468740 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468745 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468750 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468755 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468761 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468766 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468771 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468776 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468782 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468787 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468794 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468800 4983 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468823 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468830 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468835 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468841 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468846 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468853 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468860 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468866 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468872 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468877 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468883 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468889 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468894 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468900 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468905 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468911 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468917 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468922 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468927 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468934 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468940 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468945 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468950 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468955 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468961 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468966 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468971 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468976 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.468981 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.468990 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.469172 4983 server.go:940] "Client rotation is on, will bootstrap in background" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.473749 4983 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.473890 4983 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.475333 4983 server.go:997] "Starting client certificate rotation" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.475367 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.476354 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 01:51:00.148626764 +0000 UTC Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.476477 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2104h41m27.672152696s for next certificate rotation Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.504407 4983 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.509357 4983 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.533900 4983 log.go:25] "Validated CRI v1 runtime API" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.573369 4983 log.go:25] "Validated CRI v1 image API" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.575860 4983 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.581399 4983 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-01-09-05-40-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.581454 4983 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.611233 4983 manager.go:217] Machine: {Timestamp:2025-10-01 09:09:32.606309753 +0000 UTC m=+0.595538580 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:77d2c1af-84b4-4e4f-813d-45ea11b761e6 BootID:aa9ee4eb-b268-4e5c-8a70-451c4e255816 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:86:3e:ef Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:86:3e:ef Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:42:d1:f6 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:7c:88:3f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:87:08:08 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:50:3b:67 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d6:5a:78:88:9c:a2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:da:92:c1:e2:f6:c9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.611762 4983 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.612154 4983 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.615092 4983 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.615436 4983 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.615494 4983 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.615989 4983 topology_manager.go:138] "Creating topology manager with none policy" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.616018 4983 container_manager_linux.go:303] "Creating device plugin manager" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.616594 4983 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.616666 4983 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.616963 4983 state_mem.go:36] "Initialized new in-memory state store" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.617128 4983 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.621166 4983 kubelet.go:418] "Attempting to sync node with API server" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.621292 4983 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.621425 4983 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.621462 4983 kubelet.go:324] "Adding apiserver pod source" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.621915 4983 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.627079 4983 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.628444 4983 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.628970 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.629082 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.629221 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.629273 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.631661 4983 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633698 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633745 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633765 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633783 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633846 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633869 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633888 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633917 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633938 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.633977 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.634002 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.634028 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.636395 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.637177 4983 server.go:1280] "Started kubelet" Oct 01 09:09:32 crc systemd[1]: Started Kubernetes Kubelet. Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.638671 4983 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.639977 4983 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.639886 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.643336 4983 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.643862 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.643958 4983 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.644121 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.644070 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 12:18:03.652406231 +0000 UTC Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.644195 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1467h8m31.008224468s for next certificate rotation Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.645286 4983 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.645341 4983 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.649234 4983 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.649630 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.649741 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.649892 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="200ms" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652139 4983 factory.go:153] Registering CRI-O factory Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652181 4983 factory.go:221] Registration of the crio container factory successfully Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652289 4983 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652305 4983 factory.go:55] Registering systemd factory Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652319 4983 factory.go:221] Registration of the systemd container factory successfully Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652348 4983 factory.go:103] Registering Raw factory Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.652377 4983 manager.go:1196] Started watching for new ooms in manager Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.653610 4983 manager.go:319] Starting recovery of all containers Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.654575 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a52eabfe8e071 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 09:09:32.637134961 +0000 UTC m=+0.626363798,LastTimestamp:2025-10-01 09:09:32.637134961 +0000 UTC m=+0.626363798,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.656423 4983 server.go:460] "Adding debug handlers to kubelet server" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662579 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662869 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662893 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662907 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662921 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662954 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662968 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.662981 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663040 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663056 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663070 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663082 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663094 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663109 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663124 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663298 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663338 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663353 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663366 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663456 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663473 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663486 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663497 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663525 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663537 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663551 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663587 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663621 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663635 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663647 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663660 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663675 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663707 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663719 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663731 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663744 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663756 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663767 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663778 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663790 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663819 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663833 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663846 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663890 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.663901 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.664157 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.664182 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.664204 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.664228 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.664731 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665106 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665172 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665342 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665411 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665467 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665500 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665544 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665573 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665615 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665643 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665723 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665766 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665794 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665870 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665907 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665938 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.665987 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666017 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666059 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666088 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666116 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666156 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666185 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666225 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666256 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666283 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666372 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666409 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666452 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666483 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666512 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666551 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666583 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666610 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666654 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666682 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666719 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666750 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666776 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666846 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666880 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666921 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.666978 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667029 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667068 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667099 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667137 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667169 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667202 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667241 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667271 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667308 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667336 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667368 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667486 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667552 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667592 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667649 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667695 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667726 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667764 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667842 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667880 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667919 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.667982 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668025 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668072 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668112 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668152 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668201 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668297 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668342 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668380 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668407 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668433 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668469 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668500 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668533 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668562 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668587 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668623 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668649 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668676 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668710 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668739 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668774 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668799 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668872 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668906 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668943 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.668977 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669005 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669030 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669065 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669094 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669130 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669158 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669187 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669234 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669260 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669293 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669319 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.669471 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670637 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670698 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670724 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670743 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670762 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670781 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670801 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670858 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670880 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670899 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.670920 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671026 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671044 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671063 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671082 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671103 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671133 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671152 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.671171 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673196 4983 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673276 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673313 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673361 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673393 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673422 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673455 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673487 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673525 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673560 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673616 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673648 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673676 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673706 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673733 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673758 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673777 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673797 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673859 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673887 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673910 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673931 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673951 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673974 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.673996 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674031 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674051 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674072 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674094 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674114 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674136 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674157 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674176 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674216 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674239 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674262 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674339 4983 reconstruct.go:97] "Volume reconstruction finished" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.674354 4983 reconciler.go:26] "Reconciler: start to sync state" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.690286 4983 manager.go:324] Recovery completed Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.704538 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.706663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.706696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.706707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.708503 4983 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.708598 4983 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.708691 4983 state_mem.go:36] "Initialized new in-memory state store" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.710715 4983 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.713192 4983 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.713235 4983 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.713269 4983 kubelet.go:2335] "Starting kubelet main sync loop" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.713316 4983 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 01 09:09:32 crc kubenswrapper[4983]: W1001 09:09:32.714313 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.714381 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.733584 4983 policy_none.go:49] "None policy: Start" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.734299 4983 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.734382 4983 state_mem.go:35] "Initializing new in-memory state store" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.744724 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.796834 4983 manager.go:334] "Starting Device Plugin manager" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.796890 4983 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.796905 4983 server.go:79] "Starting device plugin registration server" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.797353 4983 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.797373 4983 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.797962 4983 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.798065 4983 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.798079 4983 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.809704 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.813927 4983 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.814031 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815279 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815340 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815642 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815829 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.815871 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.817073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.817109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.817132 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818434 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818553 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818782 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.818905 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819506 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819604 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819941 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.819995 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.820562 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.820594 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.820616 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821011 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821036 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821046 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821129 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821265 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821497 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.821532 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.825392 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.826281 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.826293 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.826697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.826757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.826779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.827071 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.827119 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.828362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.828388 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.828396 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.850676 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="400ms" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876582 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876636 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876663 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876685 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876707 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876756 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876838 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876873 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876893 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876930 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.876964 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.877005 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.877036 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.877068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.877089 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.897625 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.899065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.899108 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.899145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.899180 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:32 crc kubenswrapper[4983]: E1001 09:09:32.899756 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.978267 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.978630 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.978765 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.978541 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979083 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979226 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979315 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979475 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979671 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979868 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979739 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.979969 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980053 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980174 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980214 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980249 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980288 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980251 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980321 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980353 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980390 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980364 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980451 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980528 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980476 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980498 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980610 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.980733 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:32 crc kubenswrapper[4983]: I1001 09:09:32.981600 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.100736 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.102342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.102394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.102409 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.102442 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.103128 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.145046 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.167046 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.174001 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.196757 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.200448 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-419dcb017de74fb9453a136a822ff148010efc67cd3e77d624547534e815befa WatchSource:0}: Error finding container 419dcb017de74fb9453a136a822ff148010efc67cd3e77d624547534e815befa: Status 404 returned error can't find the container with id 419dcb017de74fb9453a136a822ff148010efc67cd3e77d624547534e815befa Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.200672 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.206134 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f6e458b377f00a189b4d63182edd6b7d177bd5e9053d84f9ca9f22a6528f7274 WatchSource:0}: Error finding container f6e458b377f00a189b4d63182edd6b7d177bd5e9053d84f9ca9f22a6528f7274: Status 404 returned error can't find the container with id f6e458b377f00a189b4d63182edd6b7d177bd5e9053d84f9ca9f22a6528f7274 Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.212345 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b15733174d23ff4a81d2f6c1fb0046eaa605709fc35bf50cf62c96bedd6c2210 WatchSource:0}: Error finding container b15733174d23ff4a81d2f6c1fb0046eaa605709fc35bf50cf62c96bedd6c2210: Status 404 returned error can't find the container with id b15733174d23ff4a81d2f6c1fb0046eaa605709fc35bf50cf62c96bedd6c2210 Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.217100 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-02099c2052dfbef04bd216a7a85ac4978fb866bf3b09e03f649420341443d76e WatchSource:0}: Error finding container 02099c2052dfbef04bd216a7a85ac4978fb866bf3b09e03f649420341443d76e: Status 404 returned error can't find the container with id 02099c2052dfbef04bd216a7a85ac4978fb866bf3b09e03f649420341443d76e Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.224267 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6149d2da236d661a447f9b6e62e20acf1886496dccf2b0c8c10d538272d57c68 WatchSource:0}: Error finding container 6149d2da236d661a447f9b6e62e20acf1886496dccf2b0c8c10d538272d57c68: Status 404 returned error can't find the container with id 6149d2da236d661a447f9b6e62e20acf1886496dccf2b0c8c10d538272d57c68 Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.252000 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="800ms" Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.374439 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a52eabfe8e071 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 09:09:32.637134961 +0000 UTC m=+0.626363798,LastTimestamp:2025-10-01 09:09:32.637134961 +0000 UTC m=+0.626363798,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.503854 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.505508 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.505546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.505556 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.505585 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.506038 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.643443 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.704366 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.704471 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.718193 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6149d2da236d661a447f9b6e62e20acf1886496dccf2b0c8c10d538272d57c68"} Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.719450 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"02099c2052dfbef04bd216a7a85ac4978fb866bf3b09e03f649420341443d76e"} Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.720435 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b15733174d23ff4a81d2f6c1fb0046eaa605709fc35bf50cf62c96bedd6c2210"} Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.721886 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f6e458b377f00a189b4d63182edd6b7d177bd5e9053d84f9ca9f22a6528f7274"} Oct 01 09:09:33 crc kubenswrapper[4983]: I1001 09:09:33.723160 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"419dcb017de74fb9453a136a822ff148010efc67cd3e77d624547534e815befa"} Oct 01 09:09:33 crc kubenswrapper[4983]: W1001 09:09:33.930528 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:33 crc kubenswrapper[4983]: E1001 09:09:33.930631 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:34 crc kubenswrapper[4983]: W1001 09:09:34.030845 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:34 crc kubenswrapper[4983]: E1001 09:09:34.030960 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:34 crc kubenswrapper[4983]: E1001 09:09:34.053387 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="1.6s" Oct 01 09:09:34 crc kubenswrapper[4983]: W1001 09:09:34.202477 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:34 crc kubenswrapper[4983]: E1001 09:09:34.202586 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.306923 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.308342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.308373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.308382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.308403 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:34 crc kubenswrapper[4983]: E1001 09:09:34.308872 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.643551 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.730816 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.730863 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.730873 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.730881 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.730923 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.732958 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.732996 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.733010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.733746 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8" exitCode=0 Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.733848 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.733890 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.735353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.735392 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.735404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.738032 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5" exitCode=0 Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.738113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.738295 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.740023 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.740058 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.740072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.741711 4983 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="cfdeffab1d2ba3e3ea727e4995128678360ffbe81abaf5cdea7010c89a0fdcc7" exitCode=0 Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.741778 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"cfdeffab1d2ba3e3ea727e4995128678360ffbe81abaf5cdea7010c89a0fdcc7"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.741796 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.742485 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743293 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743367 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743678 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743706 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.743722 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.745538 4983 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec" exitCode=0 Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.745592 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec"} Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.745665 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.746439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.746473 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:34 crc kubenswrapper[4983]: I1001 09:09:34.746487 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.147992 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:35 crc kubenswrapper[4983]: W1001 09:09:35.571507 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:35 crc kubenswrapper[4983]: E1001 09:09:35.571632 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.643478 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:35 crc kubenswrapper[4983]: E1001 09:09:35.654630 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="3.2s" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.752456 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.752618 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.752576 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.752657 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.754008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.754093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.754111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.758124 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.758435 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.758604 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.758725 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.758483 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.759861 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9" exitCode=0 Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.759927 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.760501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.760562 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.760580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.761173 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.762419 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f8905c1fd53245aab804a80924a7b4c1266df6bf23891c58a1786895d4893c98"} Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.762471 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.762476 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.765413 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.765450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.765470 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.765417 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.766938 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.766964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.769053 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.769114 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.769140 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: W1001 09:09:35.814158 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Oct 01 09:09:35 crc kubenswrapper[4983]: E1001 09:09:35.814298 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.909877 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.911450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.911492 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.911503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:35 crc kubenswrapper[4983]: I1001 09:09:35.911532 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:35 crc kubenswrapper[4983]: E1001 09:09:35.912017 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.437248 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.776654 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131" exitCode=0 Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.776747 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.776754 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131"} Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.777642 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.777671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.777681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781766 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f"} Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781866 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781895 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781926 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781990 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781900 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.781935 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783240 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783275 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783250 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783321 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783288 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783651 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.783989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.784024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:36 crc kubenswrapper[4983]: I1001 09:09:36.784036 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.415086 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.421863 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793166 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0"} Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793212 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace"} Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793224 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa"} Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce"} Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793245 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250"} Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793285 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793342 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793364 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.793922 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794360 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794399 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794655 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794788 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.794830 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.795482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.795501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:37 crc kubenswrapper[4983]: I1001 09:09:37.795512 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.148549 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.148627 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.614622 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.795779 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.795794 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797312 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797388 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:38 crc kubenswrapper[4983]: I1001 09:09:38.797452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.112957 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.114511 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.114846 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.114987 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.115134 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.493751 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.494072 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.495503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.495629 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.495720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.798083 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.799185 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.799246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:39 crc kubenswrapper[4983]: I1001 09:09:39.799269 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:40 crc kubenswrapper[4983]: I1001 09:09:40.778866 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:40 crc kubenswrapper[4983]: I1001 09:09:40.779031 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:40 crc kubenswrapper[4983]: I1001 09:09:40.780206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:40 crc kubenswrapper[4983]: I1001 09:09:40.780312 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:40 crc kubenswrapper[4983]: I1001 09:09:40.780386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:42 crc kubenswrapper[4983]: I1001 09:09:42.063594 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:09:42 crc kubenswrapper[4983]: I1001 09:09:42.063886 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:42 crc kubenswrapper[4983]: I1001 09:09:42.065498 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:42 crc kubenswrapper[4983]: I1001 09:09:42.065562 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:42 crc kubenswrapper[4983]: I1001 09:09:42.065587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:42 crc kubenswrapper[4983]: E1001 09:09:42.809921 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 09:09:44 crc kubenswrapper[4983]: I1001 09:09:44.716486 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:44 crc kubenswrapper[4983]: I1001 09:09:44.716615 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:44 crc kubenswrapper[4983]: I1001 09:09:44.717845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:44 crc kubenswrapper[4983]: I1001 09:09:44.717882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:44 crc kubenswrapper[4983]: I1001 09:09:44.717894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.227743 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.227820 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.232841 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.232905 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.448095 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]log ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]etcd ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/generic-apiserver-start-informers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/priority-and-fairness-filter ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-apiextensions-informers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-apiextensions-controllers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/crd-informer-synced ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-system-namespaces-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 01 09:09:46 crc kubenswrapper[4983]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/bootstrap-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/start-kube-aggregator-informers ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-registration-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-discovery-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]autoregister-completion ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-openapi-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 01 09:09:46 crc kubenswrapper[4983]: livez check failed Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.448180 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.584377 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.584626 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.585936 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.585991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:46 crc kubenswrapper[4983]: I1001 09:09:46.586008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:48 crc kubenswrapper[4983]: I1001 09:09:48.149034 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 09:09:48 crc kubenswrapper[4983]: I1001 09:09:48.149156 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.211842 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.215205 4983 trace.go:236] Trace[325388766]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:09:39.371) (total time: 11843ms): Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[325388766]: ---"Objects listed" error: 11843ms (09:09:51.215) Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[325388766]: [11.843879745s] [11.843879745s] END Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.215236 4983 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.215361 4983 trace.go:236] Trace[928233314]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:09:37.293) (total time: 13921ms): Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[928233314]: ---"Objects listed" error: 13921ms (09:09:51.215) Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[928233314]: [13.921546327s] [13.921546327s] END Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.215387 4983 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.216403 4983 trace.go:236] Trace[1445862253]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:09:36.978) (total time: 14238ms): Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[1445862253]: ---"Objects listed" error: 14238ms (09:09:51.216) Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[1445862253]: [14.238284108s] [14.238284108s] END Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.216438 4983 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.217443 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.217643 4983 trace.go:236] Trace[98160027]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:09:40.361) (total time: 10856ms): Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[98160027]: ---"Objects listed" error: 10856ms (09:09:51.217) Oct 01 09:09:51 crc kubenswrapper[4983]: Trace[98160027]: [10.856362376s] [10.856362376s] END Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.217678 4983 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.230540 4983 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.264033 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.264110 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.441672 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.442228 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.442270 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.445253 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.633953 4983 apiserver.go:52] "Watching apiserver" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.637550 4983 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.638150 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-rqrqw"] Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.638715 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.638983 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.639103 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.640012 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.640243 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.641713 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.643422 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.643672 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.643691 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.644009 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.644160 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.644356 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.644745 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.645413 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.645642 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.648725 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.648728 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.648862 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.648974 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.649006 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.649154 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.649241 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.650189 4983 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.665111 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.677012 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.690628 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.701452 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.715227 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.726376 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734376 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734428 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734455 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734482 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734506 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734530 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734552 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734574 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734600 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734628 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734650 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734672 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734725 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734731 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734757 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734778 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734854 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734905 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734927 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734948 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734967 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.734985 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735008 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735028 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735045 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735027 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735085 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735166 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735246 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735521 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735610 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735688 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735712 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735934 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735943 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736007 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.735969 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736096 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736120 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736131 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736192 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736285 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736388 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736461 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736475 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736541 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736587 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736596 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736599 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736633 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736682 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736719 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737307 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737346 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737383 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737403 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737421 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737443 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737464 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737486 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737507 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.736880 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737120 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737222 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737230 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737613 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737829 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.737891 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738026 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738035 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738034 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738067 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738116 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738138 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738154 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738172 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738188 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738197 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738205 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738299 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738320 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738402 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738428 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738452 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738481 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738501 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738521 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738540 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738559 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738578 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738596 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738614 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738634 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738652 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738634 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738697 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.738997 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739039 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739192 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739266 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739297 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739319 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739337 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.739379 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:09:52.239342137 +0000 UTC m=+20.228571134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739427 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739477 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739536 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739576 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739612 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739647 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739682 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739724 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739764 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739799 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739874 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739909 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739948 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739982 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740017 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740053 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740088 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740299 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740345 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740380 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740416 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740449 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740497 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740537 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740575 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740611 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740642 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740674 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740712 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740748 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739485 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740782 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739588 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740873 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740912 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739689 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740948 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740989 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741026 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741063 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741100 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741144 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741182 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741215 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741265 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741303 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741342 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741424 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741462 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741494 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741528 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741562 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741614 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741656 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741694 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741726 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741758 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741790 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741855 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741889 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741920 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741953 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741988 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742028 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742108 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742150 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742187 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742226 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742263 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742304 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742339 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742370 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742407 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742444 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742484 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742522 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742567 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742608 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742645 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742681 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742715 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742747 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742788 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742855 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742893 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742928 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742962 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743008 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743049 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743086 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743162 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743217 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743251 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743286 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743320 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743651 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743729 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743787 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743871 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743911 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743958 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743997 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744038 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744078 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744114 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744157 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744236 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744283 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744320 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744354 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744394 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744429 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744467 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744501 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744539 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744585 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744631 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744667 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744709 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744745 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744784 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744872 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744915 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744968 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745008 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745064 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745102 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745152 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745189 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745241 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745283 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745321 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745473 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745530 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745600 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zvd5\" (UniqueName: \"kubernetes.io/projected/9881be0c-c245-4096-bbd7-02081e1d73b9-kube-api-access-9zvd5\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745646 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745695 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745740 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745782 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745857 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745901 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9881be0c-c245-4096-bbd7-02081e1d73b9-hosts-file\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745956 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746011 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746062 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746114 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746170 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746217 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746698 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746729 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746746 4983 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746800 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746827 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746840 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746852 4983 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746867 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746879 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746890 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746903 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746914 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746926 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746936 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746948 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746958 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746977 4983 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746991 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747001 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747011 4983 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747020 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747033 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747044 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747055 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747066 4983 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747076 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747085 4983 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747096 4983 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747108 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747118 4983 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747130 4983 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747140 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747151 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747187 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747197 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747207 4983 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747217 4983 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747228 4983 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747238 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747269 4983 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747281 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747242 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747299 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745227 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.749164 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.749936 4983 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.752009 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739675 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739784 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739798 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739797 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739890 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.739215 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740100 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740218 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740287 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740346 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740358 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740441 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740966 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.740995 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741198 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741276 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741556 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741614 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741786 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741991 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742019 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742019 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742111 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.741571 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742525 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742561 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742717 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742670 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742807 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.742846 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743068 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.743709 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744438 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744478 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744499 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744515 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744764 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744786 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744879 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.744922 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745104 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745224 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745418 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745476 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745596 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745594 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745795 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745805 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745839 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745884 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745944 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745950 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.745951 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746077 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746104 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746214 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746605 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.746676 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747056 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747293 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747493 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.747607 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.759582 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:52.259528961 +0000 UTC m=+20.248757748 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747771 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747804 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.747976 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.748403 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.748475 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.748643 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.749261 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.749389 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.749560 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750075 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750093 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750165 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750246 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750261 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.750716 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751181 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751334 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.751379 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.759865 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:52.259857009 +0000 UTC m=+20.249085806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751384 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751397 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751611 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751714 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.751763 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.752922 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.756552 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.756985 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.757136 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.757611 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.762337 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.762358 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.762371 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.762422 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:52.262405253 +0000 UTC m=+20.251634050 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.767883 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.771241 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.769125 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.776104 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.776137 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.776158 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.776257 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:52.276223848 +0000 UTC m=+20.265452655 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.777168 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.777234 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.777339 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.777676 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.778406 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779017 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779416 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779504 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779574 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779738 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779829 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.779917 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.780028 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.780207 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.784442 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.785211 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.785920 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.785987 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.786421 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.786517 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.786933 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787072 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787102 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787148 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787267 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787356 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.786976 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787969 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788002 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.787186 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788271 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788328 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788399 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788484 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.788969 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.789637 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.789729 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.789992 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.790114 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.790177 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.790240 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.790290 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.790479 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.792342 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.792004 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.796084 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.796232 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.796343 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.796408 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.796473 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.798707 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.802011 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.802187 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.806453 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.806631 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.807633 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.807828 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.810009 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.810348 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.812359 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.812480 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.814962 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.817999 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.823961 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.829299 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.835269 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.846567 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.846621 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849400 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849442 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849466 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zvd5\" (UniqueName: \"kubernetes.io/projected/9881be0c-c245-4096-bbd7-02081e1d73b9-kube-api-access-9zvd5\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849502 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9881be0c-c245-4096-bbd7-02081e1d73b9-hosts-file\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849541 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849623 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849637 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849648 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.849657 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850022 4983 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850126 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850558 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9881be0c-c245-4096-bbd7-02081e1d73b9-hosts-file\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850720 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850769 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850808 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850856 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850874 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850891 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850906 4983 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850924 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850943 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850961 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850980 4983 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.850997 4983 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851013 4983 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851032 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851050 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851068 4983 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851088 4983 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851106 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851122 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851139 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851160 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851177 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851194 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851207 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851220 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851234 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851247 4983 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851261 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851314 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851351 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851363 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851377 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851391 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851404 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851417 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851429 4983 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851443 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851456 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851469 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851481 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851493 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851505 4983 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851517 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851529 4983 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851542 4983 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851556 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851569 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851582 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851594 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851608 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851622 4983 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851638 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851652 4983 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851668 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851685 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851713 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851734 4983 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851767 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851784 4983 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851797 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851842 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851860 4983 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851878 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851895 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851911 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851923 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851938 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851950 4983 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851963 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851976 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.851988 4983 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852000 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852011 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852024 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852037 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852048 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852061 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852073 4983 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852104 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852120 4983 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852134 4983 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852146 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852144 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f" exitCode=255 Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852159 4983 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852194 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852208 4983 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852223 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852233 4983 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852244 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852254 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852264 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852275 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852285 4983 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852295 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852303 4983 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852312 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852321 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852331 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852340 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852349 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852358 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852366 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852375 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852384 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852393 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852403 4983 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852415 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852424 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852433 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852442 4983 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852451 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852460 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852469 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852478 4983 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852486 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852495 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852504 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852514 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852523 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852533 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852542 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852551 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852560 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852569 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852578 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852587 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852597 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852610 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852620 4983 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852628 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852637 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852648 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852656 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852665 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852674 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852682 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852692 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852701 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852711 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852720 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852728 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852737 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852746 4983 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852754 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852764 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852773 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852782 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852790 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852800 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.852841 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f"} Oct 01 09:09:51 crc kubenswrapper[4983]: E1001 09:09:51.858499 4983 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.858803 4983 scope.go:117] "RemoveContainer" containerID="21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.868709 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.870219 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zvd5\" (UniqueName: \"kubernetes.io/projected/9881be0c-c245-4096-bbd7-02081e1d73b9-kube-api-access-9zvd5\") pod \"node-resolver-rqrqw\" (UID: \"9881be0c-c245-4096-bbd7-02081e1d73b9\") " pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.881235 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.899743 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.911659 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.922155 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.935121 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.948392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.956661 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.956754 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.967842 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:09:51 crc kubenswrapper[4983]: W1001 09:09:51.968585 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-3dd386f966cb7e51b10a6c56b5193f1f1fa8ec500f941502685ff73a17bcd57e WatchSource:0}: Error finding container 3dd386f966cb7e51b10a6c56b5193f1f1fa8ec500f941502685ff73a17bcd57e: Status 404 returned error can't find the container with id 3dd386f966cb7e51b10a6c56b5193f1f1fa8ec500f941502685ff73a17bcd57e Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.976418 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:09:51 crc kubenswrapper[4983]: I1001 09:09:51.987025 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rqrqw" Oct 01 09:09:52 crc kubenswrapper[4983]: W1001 09:09:52.100619 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9881be0c_c245_4096_bbd7_02081e1d73b9.slice/crio-3392e7c7929c8b68ab091eefca3e7e9ce54aa65d6aa26e6c8b53a249a3f86f64 WatchSource:0}: Error finding container 3392e7c7929c8b68ab091eefca3e7e9ce54aa65d6aa26e6c8b53a249a3f86f64: Status 404 returned error can't find the container with id 3392e7c7929c8b68ab091eefca3e7e9ce54aa65d6aa26e6c8b53a249a3f86f64 Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.256229 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.256370 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:09:53.256355594 +0000 UTC m=+21.245584381 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.357476 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.357514 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.357532 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.357604 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357648 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357673 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357690 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357737 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:53.357722174 +0000 UTC m=+21.346950971 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357744 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357764 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357776 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357776 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357820 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:53.357797986 +0000 UTC m=+21.347026773 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357835 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:53.357827536 +0000 UTC m=+21.347056333 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.357898 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: E1001 09:09:52.358023 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:53.35798009 +0000 UTC m=+21.347208967 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.718865 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.719687 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.721000 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.721697 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.722807 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.723420 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.724068 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.725212 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.726041 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.727202 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.727801 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.729176 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.729861 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.730274 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.730490 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.731626 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.732295 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.733430 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.733909 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.734540 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.735712 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.736208 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.737264 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.737683 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.738888 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.739003 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.739294 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.739979 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.741144 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.741603 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.742659 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.743158 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.744164 4983 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.744268 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.746120 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.747365 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.747884 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.749776 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.750568 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.751587 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.752365 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.754080 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.754646 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.755870 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.756586 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.758869 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.759562 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.760729 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.761348 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.762663 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.763279 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.764504 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.765133 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.767245 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.771714 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.772334 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.796591 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.831255 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.856826 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.859477 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.860977 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.861003 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3dd386f966cb7e51b10a6c56b5193f1f1fa8ec500f941502685ff73a17bcd57e"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.863476 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rqrqw" event={"ID":"9881be0c-c245-4096-bbd7-02081e1d73b9","Type":"ContainerStarted","Data":"f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.863506 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rqrqw" event={"ID":"9881be0c-c245-4096-bbd7-02081e1d73b9","Type":"ContainerStarted","Data":"3392e7c7929c8b68ab091eefca3e7e9ce54aa65d6aa26e6c8b53a249a3f86f64"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.864513 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d5222cc648dc23665ed98363227f593754e75f7ef421f7b25fb601fd825ab16d"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.866256 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.866348 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.866420 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d8d19b52bffa532f6dcf6a8fa5a777c259ca37667d30991031607fe0f94071b8"} Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.867897 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.896628 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.911785 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.924594 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.940410 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.957774 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.971859 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:52 crc kubenswrapper[4983]: I1001 09:09:52.987167 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.005167 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.016272 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-p92ll"] Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.016740 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-pd8zz"] Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.016855 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.017307 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.017582 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019256 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019344 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019647 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019751 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019866 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019986 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019883 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019904 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.019910 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.021388 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.024414 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mcdn7"] Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.024995 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.026916 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.026964 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.045907 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.057958 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060253 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-etc-kubernetes\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060284 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060305 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-os-release\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060325 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-hostroot\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060405 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-multus-certs\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060434 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5ckk\" (UniqueName: \"kubernetes.io/projected/d4affe98-5451-464f-af7e-6a43e5841e02-kube-api-access-d5ckk\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060467 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d4affe98-5451-464f-af7e-6a43e5841e02-mcd-auth-proxy-config\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060485 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-multus\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060508 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-system-cni-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060529 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cnibin\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060558 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-os-release\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060585 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-cni-binary-copy\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060605 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d4affe98-5451-464f-af7e-6a43e5841e02-rootfs\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060625 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060644 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-netns\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060663 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060681 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-kubelet\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060703 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-conf-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060724 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-bin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060741 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-cnibin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060761 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d4affe98-5451-464f-af7e-6a43e5841e02-proxy-tls\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060779 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060799 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vfm8\" (UniqueName: \"kubernetes.io/projected/fbf7bbd5-e570-44a5-b21a-636d54e2875a-kube-api-access-4vfm8\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060845 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-879lb\" (UniqueName: \"kubernetes.io/projected/eb6d10ad-935a-496d-8c25-32e024026cd9-kube-api-access-879lb\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060866 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-system-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060899 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-socket-dir-parent\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060920 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-k8s-cni-cncf-io\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.060941 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-daemon-config\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.069247 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.080160 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.090452 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.102654 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.113328 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.124879 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.137133 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.148978 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.160260 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161694 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-netns\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161744 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161767 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-bin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161793 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-kubelet\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161830 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-conf-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161858 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-cnibin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161880 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d4affe98-5451-464f-af7e-6a43e5841e02-proxy-tls\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161935 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-879lb\" (UniqueName: \"kubernetes.io/projected/eb6d10ad-935a-496d-8c25-32e024026cd9-kube-api-access-879lb\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161961 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.161984 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vfm8\" (UniqueName: \"kubernetes.io/projected/fbf7bbd5-e570-44a5-b21a-636d54e2875a-kube-api-access-4vfm8\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162017 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-system-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162039 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-socket-dir-parent\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162061 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-k8s-cni-cncf-io\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162414 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-daemon-config\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162436 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-cnibin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162459 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-etc-kubernetes\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162499 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162516 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-bin\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162551 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-netns\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162553 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-os-release\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162581 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-hostroot\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162604 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-socket-dir-parent\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162666 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-multus-certs\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162612 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-multus-certs\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162700 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-system-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162726 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5ckk\" (UniqueName: \"kubernetes.io/projected/d4affe98-5451-464f-af7e-6a43e5841e02-kube-api-access-d5ckk\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162783 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-kubelet\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.162846 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-conf-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163183 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-etc-kubernetes\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163208 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163229 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-run-k8s-cni-cncf-io\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163248 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d4affe98-5451-464f-af7e-6a43e5841e02-mcd-auth-proxy-config\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163277 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-multus\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163299 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-system-cni-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163331 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cnibin\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163363 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-os-release\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163379 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163399 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-cni-binary-copy\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163418 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d4affe98-5451-464f-af7e-6a43e5841e02-rootfs\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d4affe98-5451-464f-af7e-6a43e5841e02-rootfs\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163527 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163606 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cnibin\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163644 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-host-var-lib-cni-multus\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163685 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-system-cni-dir\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163764 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fbf7bbd5-e570-44a5-b21a-636d54e2875a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163857 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-hostroot\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163935 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-os-release\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163952 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-daemon-config\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.163943 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6d10ad-935a-496d-8c25-32e024026cd9-multus-cni-dir\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.164037 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d4affe98-5451-464f-af7e-6a43e5841e02-mcd-auth-proxy-config\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.164431 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6d10ad-935a-496d-8c25-32e024026cd9-cni-binary-copy\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.165743 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fbf7bbd5-e570-44a5-b21a-636d54e2875a-os-release\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.176252 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d4affe98-5451-464f-af7e-6a43e5841e02-proxy-tls\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.180919 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vfm8\" (UniqueName: \"kubernetes.io/projected/fbf7bbd5-e570-44a5-b21a-636d54e2875a-kube-api-access-4vfm8\") pod \"multus-additional-cni-plugins-mcdn7\" (UID: \"fbf7bbd5-e570-44a5-b21a-636d54e2875a\") " pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.181716 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.182126 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5ckk\" (UniqueName: \"kubernetes.io/projected/d4affe98-5451-464f-af7e-6a43e5841e02-kube-api-access-d5ckk\") pod \"machine-config-daemon-pd8zz\" (UID: \"d4affe98-5451-464f-af7e-6a43e5841e02\") " pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.188324 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-879lb\" (UniqueName: \"kubernetes.io/projected/eb6d10ad-935a-496d-8c25-32e024026cd9-kube-api-access-879lb\") pod \"multus-p92ll\" (UID: \"eb6d10ad-935a-496d-8c25-32e024026cd9\") " pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.193820 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.264435 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.264629 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:09:55.264602623 +0000 UTC m=+23.253831420 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.328547 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.335480 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p92ll" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.343380 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" Oct 01 09:09:53 crc kubenswrapper[4983]: W1001 09:09:53.351328 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb6d10ad_935a_496d_8c25_32e024026cd9.slice/crio-b38c96f1c85eb3a5eb379604a3c0eb3fcc63c62c76284587d2c2c2e7d9e33546 WatchSource:0}: Error finding container b38c96f1c85eb3a5eb379604a3c0eb3fcc63c62c76284587d2c2c2e7d9e33546: Status 404 returned error can't find the container with id b38c96f1c85eb3a5eb379604a3c0eb3fcc63c62c76284587d2c2c2e7d9e33546 Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.364879 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.364920 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.364943 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.364963 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365070 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365086 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365088 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365091 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365136 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365151 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365143 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365180 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:55.365152933 +0000 UTC m=+23.354381730 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365217 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:55.365197234 +0000 UTC m=+23.354426041 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365096 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365264 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:55.365234405 +0000 UTC m=+23.354463232 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.365310 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:55.365290946 +0000 UTC m=+23.354519863 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.383053 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7skk"] Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.384668 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.386782 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.387348 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.387891 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.388347 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.389192 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.389582 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.389767 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.406166 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.418229 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.429406 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.440241 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.456904 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.465846 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466001 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpzmm\" (UniqueName: \"kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466036 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466061 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466161 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466193 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466263 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466301 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466318 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466349 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466481 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466524 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466545 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466599 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466616 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466635 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466654 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466673 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466688 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.466716 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.474850 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.487919 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.502361 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.524875 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.544849 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567695 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567746 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567766 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpzmm\" (UniqueName: \"kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567785 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567799 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567835 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567856 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567867 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.567879 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568162 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568203 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568227 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568338 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568365 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570246 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.568385 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570325 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570383 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570446 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570478 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570519 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570549 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570630 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570688 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.570726 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571307 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571378 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571401 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571415 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571459 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571493 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571536 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.571651 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.572056 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.572115 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.572213 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.572220 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.576433 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.581369 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.586453 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpzmm\" (UniqueName: \"kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm\") pod \"ovnkube-node-f7skk\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.594826 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.713887 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.713901 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.714005 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.713916 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.714089 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:09:53 crc kubenswrapper[4983]: E1001 09:09:53.714438 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.732223 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:53 crc kubenswrapper[4983]: W1001 09:09:53.745323 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e374b0f_544c_458e_9de6_5dd3149c3dd0.slice/crio-c567ee932754affeaa92f13ca0fbdf29c244b3ce5e03efb966aa7c42d67333ae WatchSource:0}: Error finding container c567ee932754affeaa92f13ca0fbdf29c244b3ce5e03efb966aa7c42d67333ae: Status 404 returned error can't find the container with id c567ee932754affeaa92f13ca0fbdf29c244b3ce5e03efb966aa7c42d67333ae Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.869778 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" exitCode=0 Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.869843 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.870166 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"c567ee932754affeaa92f13ca0fbdf29c244b3ce5e03efb966aa7c42d67333ae"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.871831 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141" exitCode=0 Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.871890 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.871917 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerStarted","Data":"40b4a0a79fe10ec989bdd83b687858abf94c1dffb13f3820cdd71b2fb4dc5fce"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.873662 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerStarted","Data":"b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.873691 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerStarted","Data":"b38c96f1c85eb3a5eb379604a3c0eb3fcc63c62c76284587d2c2c2e7d9e33546"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.880676 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.880711 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.880720 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"1e673fc9a16346c659b60f2e5aa4db0d692862f119c4e11ea645379d03fe78ff"} Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.881236 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.887114 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.898017 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.909072 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.921317 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.940997 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.956193 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.968051 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.978257 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:53 crc kubenswrapper[4983]: I1001 09:09:53.991451 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:53Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.006970 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.017973 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.029821 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.042839 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.053197 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.062728 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.076646 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.086545 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.097173 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.115982 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.138453 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.152567 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.172931 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.183165 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.194203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.885429 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8" exitCode=0 Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.885502 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.888694 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.895966 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.896018 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.896037 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.896055 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.896073 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.907600 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.923159 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2jzdq"] Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.923477 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.927007 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.927275 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.927485 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.927656 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.928941 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.937790 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.956327 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.969285 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.983179 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/47732399-5753-4824-b164-525265f61461-serviceca\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.983439 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47732399-5753-4824-b164-525265f61461-host\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.983526 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md9b6\" (UniqueName: \"kubernetes.io/projected/47732399-5753-4824-b164-525265f61461-kube-api-access-md9b6\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.984121 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:54 crc kubenswrapper[4983]: I1001 09:09:54.998111 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.017214 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.031584 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.045925 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.059601 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.071927 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.084149 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md9b6\" (UniqueName: \"kubernetes.io/projected/47732399-5753-4824-b164-525265f61461-kube-api-access-md9b6\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.084185 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/47732399-5753-4824-b164-525265f61461-serviceca\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.084212 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47732399-5753-4824-b164-525265f61461-host\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.084465 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47732399-5753-4824-b164-525265f61461-host\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.086299 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/47732399-5753-4824-b164-525265f61461-serviceca\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.091430 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.103774 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.104416 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md9b6\" (UniqueName: \"kubernetes.io/projected/47732399-5753-4824-b164-525265f61461-kube-api-access-md9b6\") pod \"node-ca-2jzdq\" (UID: \"47732399-5753-4824-b164-525265f61461\") " pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.115942 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.128625 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.142572 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.151344 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.153674 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.154527 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.157404 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.165003 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.174103 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.186467 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.212577 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.232525 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.255564 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.267379 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.281344 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.286352 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.286552 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:09:59.286523147 +0000 UTC m=+27.275751944 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.291364 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.300496 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.303919 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2jzdq" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.311448 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: W1001 09:09:55.315947 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47732399_5753_4824_b164_525265f61461.slice/crio-ad4374281db9b11dc13b05beada08468ed0f6f73f13d98db2746fe1fd52b205f WatchSource:0}: Error finding container ad4374281db9b11dc13b05beada08468ed0f6f73f13d98db2746fe1fd52b205f: Status 404 returned error can't find the container with id ad4374281db9b11dc13b05beada08468ed0f6f73f13d98db2746fe1fd52b205f Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.322658 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.333531 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.347970 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.363124 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.374056 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.384193 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.387251 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.387288 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.387319 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.387341 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387450 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387465 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387475 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387514 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:59.387500217 +0000 UTC m=+27.376729004 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387702 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387749 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387755 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:59.387742953 +0000 UTC m=+27.376971750 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387714 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387804 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387882 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387893 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:59.387882867 +0000 UTC m=+27.377111664 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.387927 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:09:59.387911917 +0000 UTC m=+27.377140714 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.394320 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.405080 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.416727 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.430380 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.713941 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.714007 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.713943 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.714104 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.714220 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:09:55 crc kubenswrapper[4983]: E1001 09:09:55.714305 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.900266 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jzdq" event={"ID":"47732399-5753-4824-b164-525265f61461","Type":"ContainerStarted","Data":"46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe"} Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.901209 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2jzdq" event={"ID":"47732399-5753-4824-b164-525265f61461","Type":"ContainerStarted","Data":"ad4374281db9b11dc13b05beada08468ed0f6f73f13d98db2746fe1fd52b205f"} Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.905446 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.908105 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06" exitCode=0 Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.908193 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06"} Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.916726 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.931557 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.952735 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.974625 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:55 crc kubenswrapper[4983]: I1001 09:09:55.994699 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.023714 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.050278 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.062445 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.073691 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.084317 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.095424 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.107501 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.121079 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.131784 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.143167 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.155743 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.167261 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.181327 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.190505 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.200679 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.226307 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.266215 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.304138 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.346004 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.386680 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.428916 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.479072 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.504018 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.613790 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.626904 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.628842 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.629305 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.646793 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.656694 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.684291 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.724269 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.766463 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.806265 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.846561 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.884547 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.913513 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101" exitCode=0 Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.914113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101"} Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.928284 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:56 crc kubenswrapper[4983]: I1001 09:09:56.966835 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.004441 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.043633 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.084190 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.123323 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.164458 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.211165 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.248475 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.286618 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.327878 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.408190 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.428960 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.443955 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.486101 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.522183 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.562346 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.602905 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.617529 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.619600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.619640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.619652 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.619769 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.652104 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.697453 4983 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.697634 4983 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.698670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.698694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.698713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.698726 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.698734 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.713991 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.714042 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.714093 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.714135 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.714016 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.714595 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.717715 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.722316 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.722365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.722382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.722406 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.722423 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.726096 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.736139 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.742548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.742626 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.742654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.742691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.742718 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.761757 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.766975 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.767014 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.767026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.767044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.767059 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.784782 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.788438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.788474 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.788483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.788499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.788508 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.807021 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: E1001 09:09:57.807172 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.808557 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.808620 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.808631 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.808649 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.808660 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.911376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.911409 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.911420 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.911436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.911448 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:57Z","lastTransitionTime":"2025-10-01T09:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.923878 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.927515 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f" exitCode=0 Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.927608 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f"} Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.947376 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.963496 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.976930 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:57 crc kubenswrapper[4983]: I1001 09:09:57.990297 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.001497 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.013600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.013658 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.013677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.013699 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.013716 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.016254 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.037952 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.054340 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.082592 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.118222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.118270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.118287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.118307 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.118325 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.127921 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.169358 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.204963 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.221110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.221153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.221164 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.221182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.221195 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.244589 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.287612 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.323432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.323475 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.323486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.323501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.323512 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.326457 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.425663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.425715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.425732 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.425755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.425772 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.527783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.527841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.527853 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.527867 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.527879 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.630265 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.630308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.630320 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.630334 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.630345 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.732621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.732676 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.732693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.732716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.732734 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.835523 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.835558 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.835569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.835583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.835594 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.937187 4983 generic.go:334] "Generic (PLEG): container finished" podID="fbf7bbd5-e570-44a5-b21a-636d54e2875a" containerID="ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a" exitCode=0 Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.937897 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerDied","Data":"ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.937937 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.937981 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.938015 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.938043 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.938064 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:58Z","lastTransitionTime":"2025-10-01T09:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.962905 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.978678 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:58 crc kubenswrapper[4983]: I1001 09:09:58.999411 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.011218 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.024672 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.039583 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.041077 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.041129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.041141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.041158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.041172 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.053151 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.066345 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.077453 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.092695 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.115475 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.129085 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.143046 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.143077 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.143085 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.143100 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.143109 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.144115 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.156317 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.167106 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.245503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.245583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.245607 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.245642 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.245665 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.325094 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.325302 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.325284508 +0000 UTC m=+35.314513305 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.349264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.349617 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.349860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.350207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.350413 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.426135 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.426181 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.426216 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.426249 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426370 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426386 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426399 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426409 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426448 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.426432734 +0000 UTC m=+35.415661531 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426477 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426493 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.426469065 +0000 UTC m=+35.415697882 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426512 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426564 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426595 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426569 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.426542336 +0000 UTC m=+35.415771213 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.426736 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.42669737 +0000 UTC m=+35.415926247 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.452550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.452595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.452608 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.452629 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.452640 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.555206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.555281 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.555306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.555342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.555368 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.658480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.658543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.658562 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.658593 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.658615 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.713856 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.713892 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.713987 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.714101 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.714317 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:09:59 crc kubenswrapper[4983]: E1001 09:09:59.714500 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.761485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.761528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.761540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.761560 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.761572 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.864729 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.864787 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.864831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.864864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.864920 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.946911 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.947296 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.952789 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" event={"ID":"fbf7bbd5-e570-44a5-b21a-636d54e2875a","Type":"ContainerStarted","Data":"26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.967120 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.967168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.967183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.967204 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.967235 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:09:59Z","lastTransitionTime":"2025-10-01T09:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.976169 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.976593 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:09:59 crc kubenswrapper[4983]: I1001 09:09:59.993722 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:09:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.010027 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.026737 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.043321 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.066972 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.070864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.070920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.070940 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.070967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.070986 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.089941 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.110940 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.130943 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.148618 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.168837 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.173307 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.173361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.173380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.173407 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.173426 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.201258 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.218053 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.236687 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.255893 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.271357 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.276708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.276777 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.276844 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.276880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.276904 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.290873 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.308203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.337561 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.358983 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.372355 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.379431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.379469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.379480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.379495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.379503 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.387748 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.402040 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.418937 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.434858 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.449734 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.461192 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.477238 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.482763 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.482856 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.482871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.482888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.482923 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.500850 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.513824 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.585378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.585435 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.585445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.585461 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.585471 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.688576 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.688910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.689014 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.689106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.689195 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.792347 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.792404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.792417 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.792438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.792455 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.895199 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.895243 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.895257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.895283 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.895295 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.956281 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.956968 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.986369 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.998145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.998177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.998188 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.998202 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:00 crc kubenswrapper[4983]: I1001 09:10:00.998213 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:00Z","lastTransitionTime":"2025-10-01T09:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.011679 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.044385 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.064867 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.084221 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.095915 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.102520 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.102654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.102672 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.102692 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.102738 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.127224 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.149583 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.164999 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.180016 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.194036 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.205835 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.205921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.205939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.205961 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.205978 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.213622 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.231578 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.246649 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.263534 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.278270 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:01Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.308398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.308464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.308506 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.308532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.308549 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.411795 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.411882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.411900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.411922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.411937 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.515547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.515588 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.515600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.515618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.515629 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.618845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.618891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.618906 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.618924 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.618936 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.714223 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.714266 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.714324 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:01 crc kubenswrapper[4983]: E1001 09:10:01.714343 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:01 crc kubenswrapper[4983]: E1001 09:10:01.714498 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:01 crc kubenswrapper[4983]: E1001 09:10:01.714607 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.720854 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.720889 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.720899 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.720913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.720924 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.823096 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.823132 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.823142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.823158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.823169 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.925092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.925132 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.925144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.925163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:01 crc kubenswrapper[4983]: I1001 09:10:01.925174 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:01Z","lastTransitionTime":"2025-10-01T09:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.028418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.028480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.028497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.028528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.028545 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.131182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.131234 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.131246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.131264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.131276 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.233429 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.233474 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.233489 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.233525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.233538 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.336098 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.336147 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.336161 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.336205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.336219 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.438725 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.438760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.438769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.438783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.438792 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.540853 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.540920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.540938 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.540967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.540985 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.643713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.643786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.643858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.643889 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.643906 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.737188 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.746608 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.746746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.746832 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.746900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.746958 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.793737 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.810253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.833092 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.849354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.849391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.849399 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.849418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.849452 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.852224 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.865548 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.878352 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.897125 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.924351 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.940883 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.951660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.951709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.951721 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.951746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.951762 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:02Z","lastTransitionTime":"2025-10-01T09:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.955937 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.962463 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/0.log" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.965230 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0" exitCode=1 Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.965289 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0"} Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.966366 4983 scope.go:117] "RemoveContainer" containerID="49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.973567 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.985135 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:02 crc kubenswrapper[4983]: I1001 09:10:02.999064 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:02Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.012756 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.025506 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.037646 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.048117 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.055523 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.055750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.055879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.056001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.056085 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.062370 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.075357 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.086266 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.096837 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.116077 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.137002 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.149517 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.158771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.158814 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.158823 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.158838 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.158848 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.161763 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.171514 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.187054 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.231235 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:02Z\\\",\\\"message\\\":\\\"l event handler 9\\\\nI1001 09:10:02.350373 6316 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:02.352643 6316 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.352900 6316 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353065 6316 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353199 6316 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353217 6316 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353758 6316 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:02.354035 6316 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:02.354058 6316 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:02.354124 6316 factory.go:656] Stopping watch factory\\\\nI1001 09:10:02.354148 6316 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:02.354198 6316 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:02.354213 6316 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.258648 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.261218 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.261238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.261247 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.261260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.261268 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.363852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.363893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.363905 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.363923 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.363935 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.466408 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.466653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.466671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.466687 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.466697 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.568886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.568917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.568926 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.568937 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.568945 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.670874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.670923 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.670945 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.670971 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.670991 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.714588 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.714609 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.714708 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:03 crc kubenswrapper[4983]: E1001 09:10:03.714940 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:03 crc kubenswrapper[4983]: E1001 09:10:03.715033 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:03 crc kubenswrapper[4983]: E1001 09:10:03.715106 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.773671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.773717 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.773736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.773760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.773778 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.876118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.876165 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.876177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.876193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.876205 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.970225 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/1.log" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.970999 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/0.log" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.974563 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0" exitCode=1 Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.974619 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0"} Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.974683 4983 scope.go:117] "RemoveContainer" containerID="49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.975632 4983 scope.go:117] "RemoveContainer" containerID="aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0" Oct 01 09:10:03 crc kubenswrapper[4983]: E1001 09:10:03.975906 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.979098 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.979152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.979168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.979193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:03 crc kubenswrapper[4983]: I1001 09:10:03.979210 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:03Z","lastTransitionTime":"2025-10-01T09:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.001113 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:03Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.015492 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.033987 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.051092 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.080392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.081503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.081576 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.081595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.081621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.081640 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.102870 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:02Z\\\",\\\"message\\\":\\\"l event handler 9\\\\nI1001 09:10:02.350373 6316 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:02.352643 6316 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.352900 6316 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353065 6316 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353199 6316 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353217 6316 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353758 6316 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:02.354035 6316 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:02.354058 6316 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:02.354124 6316 factory.go:656] Stopping watch factory\\\\nI1001 09:10:02.354148 6316 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:02.354198 6316 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:02.354213 6316 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.119655 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.134412 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.154312 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.167041 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.184951 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.185042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.185066 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.185102 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.185126 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.185985 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.198615 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.209099 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.218906 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.238139 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.284605 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.287785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.287872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.287890 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.287913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.287934 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.300111 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.317340 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.327042 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.357385 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.372900 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.388660 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.390099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.390153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.390166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.390182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.390192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.401403 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.419609 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.439416 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.458105 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.472767 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.489315 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.493464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.493520 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.493545 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.493577 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.493600 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.506556 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.527160 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:02Z\\\",\\\"message\\\":\\\"l event handler 9\\\\nI1001 09:10:02.350373 6316 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:02.352643 6316 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.352900 6316 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353065 6316 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353199 6316 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353217 6316 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353758 6316 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:02.354035 6316 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:02.354058 6316 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:02.354124 6316 factory.go:656] Stopping watch factory\\\\nI1001 09:10:02.354148 6316 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:02.354198 6316 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:02.354213 6316 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.541392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.596493 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.596555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.596567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.596584 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.596598 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.655524 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8"] Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.656064 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.659694 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.662088 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.674946 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.690037 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.699436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.699482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.699496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.699513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.699524 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.706934 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.713484 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.713540 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.713617 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.713650 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh4j5\" (UniqueName: \"kubernetes.io/projected/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-kube-api-access-rh4j5\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.723927 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.753461 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.767044 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.777441 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.787755 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.801299 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.802000 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.802033 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.802045 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.802062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.802074 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.814088 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.814282 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh4j5\" (UniqueName: \"kubernetes.io/projected/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-kube-api-access-rh4j5\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.814349 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.814374 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.814419 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.815083 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.815356 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.824315 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.828927 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.833390 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh4j5\" (UniqueName: \"kubernetes.io/projected/3bf4a18e-a1a1-45e4-a577-e0442040b6c3-kube-api-access-rh4j5\") pod \"ovnkube-control-plane-749d76644c-4jlf8\" (UID: \"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.843372 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.856260 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.868291 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.882347 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.905551 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49c2e90195c145a2144c499236f3d94992284e97efdbba8b7cc5081053bd3eb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:02Z\\\",\\\"message\\\":\\\"l event handler 9\\\\nI1001 09:10:02.350373 6316 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:02.352643 6316 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.352900 6316 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353065 6316 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353199 6316 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353217 6316 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:10:02.353758 6316 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:02.354035 6316 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:02.354058 6316 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:02.354124 6316 factory.go:656] Stopping watch factory\\\\nI1001 09:10:02.354148 6316 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:02.354198 6316 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:02.354213 6316 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.906102 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.906135 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.906149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.906168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.906179 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:04Z","lastTransitionTime":"2025-10-01T09:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.974125 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.979422 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/1.log" Oct 01 09:10:04 crc kubenswrapper[4983]: I1001 09:10:04.983834 4983 scope.go:117] "RemoveContainer" containerID="aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0" Oct 01 09:10:04 crc kubenswrapper[4983]: E1001 09:10:04.984059 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.003544 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.013002 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.013045 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.013057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.013076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.013088 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.020654 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.040359 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.054602 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.066391 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.081184 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.093872 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.116142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.116182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.116196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.116214 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.116226 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.123691 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.139957 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.153137 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.170018 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.191673 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.206717 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.220003 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.220049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.220216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.220244 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.220273 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.223511 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.236242 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.248848 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:05Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.323003 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.323042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.323053 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.323120 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.323139 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.426177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.426239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.426257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.426284 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.426302 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.529556 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.529609 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.529629 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.529654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.529673 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.632028 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.632072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.632080 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.632097 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.632107 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.713968 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.714125 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:05 crc kubenswrapper[4983]: E1001 09:10:05.714305 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.714401 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:05 crc kubenswrapper[4983]: E1001 09:10:05.714480 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:05 crc kubenswrapper[4983]: E1001 09:10:05.714608 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.734032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.734064 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.734075 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.734092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.734103 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.837199 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.837232 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.837244 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.837260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.837276 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.939923 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.939961 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.939974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.939989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.939998 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:05Z","lastTransitionTime":"2025-10-01T09:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.988482 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" event={"ID":"3bf4a18e-a1a1-45e4-a577-e0442040b6c3","Type":"ContainerStarted","Data":"a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.988537 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" event={"ID":"3bf4a18e-a1a1-45e4-a577-e0442040b6c3","Type":"ContainerStarted","Data":"b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304"} Oct 01 09:10:05 crc kubenswrapper[4983]: I1001 09:10:05.988551 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" event={"ID":"3bf4a18e-a1a1-45e4-a577-e0442040b6c3","Type":"ContainerStarted","Data":"12624d0659347fc2c40ff656f400ea73ea27e09905f24103131320c339a64b3b"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.007847 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.021383 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.032021 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.042372 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.042439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.042456 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.042479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.042497 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.048064 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.059440 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.076563 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.092193 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.121522 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.141947 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.144852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.144919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.144932 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.144950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.144962 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.154917 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.171404 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.181096 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-pvhnl"] Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.181763 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: E1001 09:10:06.181943 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.190417 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.205747 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.222482 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.240071 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.247708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.247742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.247754 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.247771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.247783 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.258398 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.278887 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.292881 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.307939 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.323742 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.331568 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bfw\" (UniqueName: \"kubernetes.io/projected/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-kube-api-access-67bfw\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.331624 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.349363 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.351914 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.351970 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.351987 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.352011 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.352030 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.381699 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.397988 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.415469 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.431929 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.432163 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bfw\" (UniqueName: \"kubernetes.io/projected/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-kube-api-access-67bfw\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.432240 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: E1001 09:10:06.432504 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:06 crc kubenswrapper[4983]: E1001 09:10:06.432650 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:06.932620154 +0000 UTC m=+34.921848991 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.449783 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454761 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.454931 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bfw\" (UniqueName: \"kubernetes.io/projected/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-kube-api-access-67bfw\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.465684 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.484221 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.509299 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.531513 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.550411 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.557218 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.557277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.557295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.557335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.557352 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.567951 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.589775 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:06Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.660365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.660406 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.660417 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.660434 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.660445 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.763327 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.763397 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.763414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.763439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.763456 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.867048 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.867109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.867126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.867153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.867172 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.938242 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:06 crc kubenswrapper[4983]: E1001 09:10:06.938545 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:06 crc kubenswrapper[4983]: E1001 09:10:06.938695 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:07.938667386 +0000 UTC m=+35.927896213 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.970242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.970296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.970315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.970341 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:06 crc kubenswrapper[4983]: I1001 09:10:06.970384 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:06Z","lastTransitionTime":"2025-10-01T09:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.072898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.072951 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.072999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.073024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.073042 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.176237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.176298 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.176315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.176342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.176361 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.279714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.280142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.280458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.280657 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.280976 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.343663 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.343935 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:10:23.343884421 +0000 UTC m=+51.333113258 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.384143 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.384220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.384239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.384266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.384289 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.445451 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.445538 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.445599 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.445641 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445737 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445795 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445896 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:23.445873897 +0000 UTC m=+51.435102734 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445896 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445985 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446044 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446071 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.445923 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446179 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:23.446146294 +0000 UTC m=+51.435375121 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446235 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:23.446209467 +0000 UTC m=+51.435438294 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446007 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.446289 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:23.446277568 +0000 UTC m=+51.435506395 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.487541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.487617 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.487641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.487670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.487689 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.592042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.592402 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.592594 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.593047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.593086 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.696720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.696784 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.696849 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.696884 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.696918 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.714353 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.714411 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.714442 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.714533 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.714366 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.714801 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.714958 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.715117 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.799945 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.800021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.800054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.800084 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.800104 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.903139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.903500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.903516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.903536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.903549 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:07Z","lastTransitionTime":"2025-10-01T09:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:07 crc kubenswrapper[4983]: I1001 09:10:07.950246 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.950450 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:07 crc kubenswrapper[4983]: E1001 09:10:07.950542 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:09.950518756 +0000 UTC m=+37.939747593 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.005886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.005942 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.005960 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.005984 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.006002 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.058876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.058920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.058928 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.058949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.058958 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.073174 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.078227 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.078287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.078306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.078338 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.078356 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.093430 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.097938 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.098000 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.098012 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.098027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.098040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.116960 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.121750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.121797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.121821 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.121837 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.121845 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.139280 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.143121 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.143196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.143220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.143247 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.143266 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.161712 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:08 crc kubenswrapper[4983]: E1001 09:10:08.161845 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.163193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.163220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.163231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.163248 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.163261 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.267083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.267168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.267189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.267226 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.267247 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.371228 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.371306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.371326 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.371354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.371375 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.474682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.474743 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.474761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.474795 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.474842 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.578429 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.578495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.578513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.578539 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.578558 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.681638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.681693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.681710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.681733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.681751 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.784414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.784801 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.785135 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.785349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.785521 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.893028 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.893083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.893101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.893127 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.893146 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.996329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.996432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.996452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.996475 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:08 crc kubenswrapper[4983]: I1001 09:10:08.996541 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:08Z","lastTransitionTime":"2025-10-01T09:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.100765 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.100902 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.100923 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.100949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.100974 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.203731 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.203793 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.203838 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.203865 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.203881 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.306916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.306979 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.306996 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.307021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.307039 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.410516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.410584 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.410601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.410627 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.410643 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.513139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.513204 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.513216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.513236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.513248 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.616292 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.616357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.616380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.616409 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.616433 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.713803 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.713927 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.713850 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.714030 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.713975 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.714152 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.714366 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.714540 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.719643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.719689 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.719708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.719730 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.719746 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.823298 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.823364 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.823382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.823410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.823430 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.926536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.926593 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.926611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.926635 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.926653 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:09Z","lastTransitionTime":"2025-10-01T09:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:09 crc kubenswrapper[4983]: I1001 09:10:09.973630 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.974146 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:09 crc kubenswrapper[4983]: E1001 09:10:09.974282 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:13.974254416 +0000 UTC m=+41.963483253 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.029510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.029604 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.029621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.029644 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.029663 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.132324 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.132394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.132417 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.132451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.132475 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.235075 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.235134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.235156 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.235185 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.235203 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.337913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.337982 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.338001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.338024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.338042 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.441296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.441353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.441370 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.441394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.441411 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.544761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.544866 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.544885 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.545317 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.545367 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.649656 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.650045 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.650279 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.650484 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.650689 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.754509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.754569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.754586 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.754610 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.754628 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.857611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.857673 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.857690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.857713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.857730 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.961260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.961336 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.961355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.961380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:10 crc kubenswrapper[4983]: I1001 09:10:10.961398 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:10Z","lastTransitionTime":"2025-10-01T09:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.064638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.064694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.064716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.064744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.064767 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.167528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.167581 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.167598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.167622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.167639 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.271032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.271107 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.271131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.271164 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.271190 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.374659 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.374721 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.374741 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.374766 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.374783 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.477606 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.477665 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.477681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.477705 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.477722 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.581156 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.581206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.581222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.581246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.581263 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.685010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.685505 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.685735 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.685917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.686072 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.713725 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.713882 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.713759 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:11 crc kubenswrapper[4983]: E1001 09:10:11.714033 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.714113 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:11 crc kubenswrapper[4983]: E1001 09:10:11.714305 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:11 crc kubenswrapper[4983]: E1001 09:10:11.714508 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:11 crc kubenswrapper[4983]: E1001 09:10:11.714677 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.788999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.789052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.789064 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.789082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.789096 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.891838 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.891912 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.891939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.891973 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.891996 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.994106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.994152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.994163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.994181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:11 crc kubenswrapper[4983]: I1001 09:10:11.994192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:11Z","lastTransitionTime":"2025-10-01T09:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.096848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.096910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.096927 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.096950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.096966 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.200021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.200078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.200096 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.200119 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.200136 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.302618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.302690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.302716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.302746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.302769 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.405488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.405540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.405557 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.405580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.405600 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.508528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.508617 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.508640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.508672 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.508694 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.611415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.611497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.611530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.611561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.611581 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.715034 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.715162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.715194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.715238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.715267 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.739114 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.758074 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.778182 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.798730 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.818488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.818592 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.818613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.818641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.818662 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.820606 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.839389 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.855789 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.869967 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.894607 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.912957 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.924563 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.924597 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.924605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.924619 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.924627 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:12Z","lastTransitionTime":"2025-10-01T09:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.947448 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:12 crc kubenswrapper[4983]: I1001 09:10:12.992098 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:12Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.014714 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:13Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.026850 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.026879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.026892 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.026909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.026921 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.037689 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:13Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.058384 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:13Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.070614 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:13Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.084204 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:13Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.129178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.129228 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.129239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.129255 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.129267 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.231898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.231943 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.231955 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.231969 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.231981 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.334681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.334726 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.334736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.334751 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.334760 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.436998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.437109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.437122 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.437138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.437153 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.540025 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.540088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.540106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.540129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.540148 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.643167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.643224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.643243 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.643266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.643284 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.713444 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.713478 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.713633 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:13 crc kubenswrapper[4983]: E1001 09:10:13.713845 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.713894 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:13 crc kubenswrapper[4983]: E1001 09:10:13.714030 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:13 crc kubenswrapper[4983]: E1001 09:10:13.714148 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:13 crc kubenswrapper[4983]: E1001 09:10:13.714252 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.746472 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.746525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.746543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.746567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.746584 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.849022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.849363 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.849515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.849665 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.849859 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.953083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.953387 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.953535 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.953681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:13 crc kubenswrapper[4983]: I1001 09:10:13.953842 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:13Z","lastTransitionTime":"2025-10-01T09:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: E1001 09:10:14.017536 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:14 crc kubenswrapper[4983]: E1001 09:10:14.017666 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:22.017634963 +0000 UTC m=+50.006863800 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.017354 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.056095 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.056161 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.056181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.056207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.056224 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.158836 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.158881 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.158896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.158917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.158932 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.262302 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.262384 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.262410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.262447 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.262471 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.366242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.366315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.366340 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.366365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.366382 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.472155 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.472735 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.472770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.472841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.472873 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.575873 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.575924 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.575943 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.575965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.575981 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.679082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.679439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.679596 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.679750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.679944 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.783263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.783315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.783331 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.783354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.783370 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.886253 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.886357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.886379 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.886403 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.886422 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.989426 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.989491 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.989510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.989536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:14 crc kubenswrapper[4983]: I1001 09:10:14.989554 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:14Z","lastTransitionTime":"2025-10-01T09:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.092996 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.093068 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.093094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.093126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.093159 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.196095 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.196176 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.196204 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.196234 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.196256 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.299643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.299710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.299727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.299752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.299771 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.403051 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.403115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.403133 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.403162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.403183 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.506092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.506149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.506168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.506190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.506207 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.609304 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.609436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.609454 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.609479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.609497 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713225 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713243 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713283 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713552 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:15 crc kubenswrapper[4983]: E1001 09:10:15.713671 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713745 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:15 crc kubenswrapper[4983]: E1001 09:10:15.713860 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.713927 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:15 crc kubenswrapper[4983]: E1001 09:10:15.714004 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.714106 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:15 crc kubenswrapper[4983]: E1001 09:10:15.714336 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.816289 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.816369 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.816391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.816421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.816439 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.919651 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.919760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.919779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.919803 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:15 crc kubenswrapper[4983]: I1001 09:10:15.919846 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:15Z","lastTransitionTime":"2025-10-01T09:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.023439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.023507 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.023530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.023561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.023587 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.126880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.126970 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.127004 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.127035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.127057 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.230044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.230105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.230126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.230150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.230168 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.333104 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.333215 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.333270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.333301 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.333320 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.436871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.436936 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.436954 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.436981 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.437002 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.541178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.541249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.541269 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.541297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.541321 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.646095 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.646181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.646205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.646237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.646262 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.750223 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.750329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.750352 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.750377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.750395 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.854658 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.854736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.854761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.854797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.854861 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.958663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.958728 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.958751 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.958790 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:16 crc kubenswrapper[4983]: I1001 09:10:16.958839 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:16Z","lastTransitionTime":"2025-10-01T09:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.062789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.062916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.062936 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.062963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.062990 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.166548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.166620 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.166640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.166674 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.166697 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.270144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.270205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.270232 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.270263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.270288 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.373921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.373984 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.374034 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.374058 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.374072 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.477483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.477567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.477585 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.477611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.477629 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.581750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.581949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.581975 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.582005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.582030 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.685596 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.685676 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.685697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.685740 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.685767 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.714433 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.714459 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.714506 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:17 crc kubenswrapper[4983]: E1001 09:10:17.714652 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.714725 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:17 crc kubenswrapper[4983]: E1001 09:10:17.715038 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:17 crc kubenswrapper[4983]: E1001 09:10:17.715124 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:17 crc kubenswrapper[4983]: E1001 09:10:17.715252 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.789588 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.789702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.789748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.789789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.790022 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.894266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.894351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.894371 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.894405 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.894429 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.999447 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.999566 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:17 crc kubenswrapper[4983]: I1001 09:10:17.999591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:17.999676 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:17.999747 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:17Z","lastTransitionTime":"2025-10-01T09:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.102727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.102779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.102799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.102852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.102872 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.206785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.206892 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.206913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.206945 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.206967 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.310089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.310172 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.310190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.310222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.310244 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.412964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.413008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.413020 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.413039 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.413051 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.455361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.455450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.455468 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.455499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.455518 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.477694 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.484179 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.484267 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.484287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.484332 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.484353 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.507955 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.513683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.513752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.513772 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.513800 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.513857 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.540210 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.546769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.546889 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.546915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.546950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.546978 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.568403 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.575444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.575532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.575554 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.575593 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.575616 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.596892 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:18 crc kubenswrapper[4983]: E1001 09:10:18.597091 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.600512 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.600566 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.600580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.600605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.600622 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.704027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.704088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.704106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.704130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.704147 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.806747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.806852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.806871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.806900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.806919 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.910622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.910697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.910733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.910763 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:18 crc kubenswrapper[4983]: I1001 09:10:18.910785 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:18Z","lastTransitionTime":"2025-10-01T09:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.013219 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.013266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.013278 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.013295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.013308 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.115868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.115937 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.115972 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.116005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.116029 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.220065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.220148 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.220173 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.220204 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.220226 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.323151 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.323202 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.323218 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.323240 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.323257 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.429394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.429948 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.430041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.430173 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.430256 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.533516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.533886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.534180 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.534416 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.534590 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.637839 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.637880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.637899 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.637919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.637935 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.718251 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.718414 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:19 crc kubenswrapper[4983]: E1001 09:10:19.718536 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.718624 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.718707 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:19 crc kubenswrapper[4983]: E1001 09:10:19.718803 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:19 crc kubenswrapper[4983]: E1001 09:10:19.719140 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:19 crc kubenswrapper[4983]: E1001 09:10:19.719361 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.720751 4983 scope.go:117] "RemoveContainer" containerID="aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.741098 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.741162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.741183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.741209 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.741227 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.844089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.844139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.844152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.844171 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.844184 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.946632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.947052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.947068 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.947088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:19 crc kubenswrapper[4983]: I1001 09:10:19.947102 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:19Z","lastTransitionTime":"2025-10-01T09:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.045348 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/1.log" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.048943 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.048985 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.048995 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.049011 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.049024 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.049904 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.050660 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.063248 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.083190 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.108183 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.123142 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.142456 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.152064 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.152111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.152126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.152144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.152158 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.160627 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.175597 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.189100 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.211759 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.229246 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.246885 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.254745 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.254922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.254989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.255066 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.255140 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.263120 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.280505 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.294203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.307911 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.318948 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.329713 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.357578 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.357621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.357632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.357650 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.357663 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.460052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.460094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.460105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.460121 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.460134 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.565999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.566044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.566054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.566076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.566091 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.668227 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.668276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.668289 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.668310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.668325 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.771050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.771088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.771097 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.771112 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.771124 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.873832 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.873877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.873887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.873902 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.873912 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.976296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.976344 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.976355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.976373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:20 crc kubenswrapper[4983]: I1001 09:10:20.976385 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:20Z","lastTransitionTime":"2025-10-01T09:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.053974 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/2.log" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.054455 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/1.log" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.056882 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" exitCode=1 Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.056935 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.056979 4983 scope.go:117] "RemoveContainer" containerID="aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.057971 4983 scope.go:117] "RemoveContainer" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" Oct 01 09:10:21 crc kubenswrapper[4983]: E1001 09:10:21.058277 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.069619 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.080326 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.080389 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.080408 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.080438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.080459 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.085915 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.098993 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.111136 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.126614 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.140603 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.169568 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.183242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.183289 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.183300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.183319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.183332 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.186764 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.199969 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.212430 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.230042 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa0932c9d7b49debdf4d8b660b9347a2d6c9456e80fef3852faa8eb14ab48fa0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:03Z\\\",\\\"message\\\":\\\"I1001 09:10:03.765479 6435 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:10:03.765513 6435 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:10:03.765526 6435 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:10:03.765562 6435 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:10:03.766351 6435 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 09:10:03.766372 6435 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:03.766354 6435 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:10:03.766408 6435 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:03.766417 6435 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:03.766434 6435 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:03.766437 6435 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:03.766455 6435 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:03.766477 6435 factory.go:656] Stopping watch factory\\\\nI1001 09:10:03.766489 6435 ovnkube.go:599] Stopped ovnkube\\\\nI1001 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.245483 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.263344 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.281128 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.285956 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.285989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.286002 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.286022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.286035 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.296428 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.311170 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.325505 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.388270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.388315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.388325 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.388339 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.388349 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.490080 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.490126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.490134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.490150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.490159 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.592244 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.592497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.592600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.592677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.592740 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.695373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.695762 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.695952 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.696089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.696199 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.714123 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.714227 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.714168 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.714129 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:21 crc kubenswrapper[4983]: E1001 09:10:21.714547 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:21 crc kubenswrapper[4983]: E1001 09:10:21.714688 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:21 crc kubenswrapper[4983]: E1001 09:10:21.714707 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:21 crc kubenswrapper[4983]: E1001 09:10:21.714747 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.798493 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.798746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.798830 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.798918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.798998 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.900645 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.900684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.900694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.900709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:21 crc kubenswrapper[4983]: I1001 09:10:21.900719 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:21Z","lastTransitionTime":"2025-10-01T09:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.003625 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.003727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.003748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.003848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.003869 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.062978 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/2.log" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.069387 4983 scope.go:117] "RemoveContainer" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.069617 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:10:22 crc kubenswrapper[4983]: E1001 09:10:22.069765 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.083702 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.085283 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.097401 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.106169 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.106216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.106228 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.106246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.106261 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.110857 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.113622 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:22 crc kubenswrapper[4983]: E1001 09:10:22.114054 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:22 crc kubenswrapper[4983]: E1001 09:10:22.114105 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:38.114088729 +0000 UTC m=+66.103317526 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.128547 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.142739 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.159079 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.173043 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.191123 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.204408 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.207951 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.208041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.208113 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.208177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.208239 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.215902 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.226629 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.236483 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.247535 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.260712 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.275385 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.288779 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.298005 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.310488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.310763 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.310891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.310983 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.311066 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.325058 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.343791 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.363631 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.378528 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.393740 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.409429 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.414236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.414266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.414275 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.414290 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.414300 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.421878 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.434028 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.446139 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.459098 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.470771 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.482999 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.492892 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.513691 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.516070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.516107 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.516116 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.516132 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.516141 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.524751 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.536348 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.548353 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.558498 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.618530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.618575 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.618589 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.618607 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.618620 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.720270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.720303 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.720311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.720323 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.720332 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.725224 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.735439 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.746051 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.754897 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.765211 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.778319 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.798240 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.821833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.821872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.821886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.821903 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.821916 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.857161 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.867928 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.879841 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.889955 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.905785 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.916981 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.924503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.924542 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.924552 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.924567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.924579 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:22Z","lastTransitionTime":"2025-10-01T09:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.934634 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.947424 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.961025 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:22 crc kubenswrapper[4983]: I1001 09:10:22.976026 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.002740 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.026733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.026773 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.026788 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.026828 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.026843 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.130384 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.130431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.130444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.130463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.130475 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.232492 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.232524 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.232532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.232546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.232554 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.335976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.336032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.336049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.336073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.336090 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.426241 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.426531 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:10:55.426506181 +0000 UTC m=+83.415735018 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.439184 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.439230 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.439248 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.439271 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.439292 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.527786 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.527936 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.527982 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.528016 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528025 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528149 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:55.528123527 +0000 UTC m=+83.517352364 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528197 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528226 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528244 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528266 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528312 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:55.528291212 +0000 UTC m=+83.517520039 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528335 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528386 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528409 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528344 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:55.528322913 +0000 UTC m=+83.517551740 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.528541 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:10:55.528500237 +0000 UTC m=+83.517729064 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.541654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.541703 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.541720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.541742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.541760 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.645082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.645120 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.645129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.645144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.645157 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.714091 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.714144 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.714222 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.714277 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.714454 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.714600 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.714767 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:23 crc kubenswrapper[4983]: E1001 09:10:23.714940 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.747752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.747829 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.747845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.747872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.747895 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.850206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.850251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.850262 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.850277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.850287 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.953129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.953166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.953174 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.953188 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:23 crc kubenswrapper[4983]: I1001 09:10:23.953201 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:23Z","lastTransitionTime":"2025-10-01T09:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.056359 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.056408 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.056460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.056481 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.056497 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.159361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.159423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.159439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.159468 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.159486 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.262545 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.262647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.262667 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.262690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.262709 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.366136 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.366169 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.366177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.366222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.366232 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.469306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.469391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.469408 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.469431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.469449 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.571615 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.571677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.571694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.571719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.571736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.675168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.675217 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.675229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.675246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.675258 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.778518 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.778567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.778587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.778609 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.778626 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.881371 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.881438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.881455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.881479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.881547 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.984316 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.984357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.984368 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.984383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:24 crc kubenswrapper[4983]: I1001 09:10:24.984393 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:24Z","lastTransitionTime":"2025-10-01T09:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.086690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.086985 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.087086 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.087176 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.087276 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.189643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.189680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.189691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.189707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.189718 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.292021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.292055 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.292066 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.292081 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.292092 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.394599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.394640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.394656 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.394679 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.394695 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.496400 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.496433 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.496444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.496460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.496471 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.599297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.599350 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.599362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.599381 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.599396 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.701352 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.701387 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.701399 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.701412 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.701420 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.713958 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.713980 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.714001 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:25 crc kubenswrapper[4983]: E1001 09:10:25.714038 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.714106 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:25 crc kubenswrapper[4983]: E1001 09:10:25.714175 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:25 crc kubenswrapper[4983]: E1001 09:10:25.714300 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:25 crc kubenswrapper[4983]: E1001 09:10:25.714330 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.804435 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.804548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.804567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.804597 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.804614 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.907836 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.907866 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.907874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.907887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:25 crc kubenswrapper[4983]: I1001 09:10:25.907897 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:25Z","lastTransitionTime":"2025-10-01T09:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.010303 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.010337 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.010345 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.010358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.010367 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.112665 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.112710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.112718 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.112731 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.112740 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.214894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.214939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.214949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.214965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.214979 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.317319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.317376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.317400 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.317431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.317455 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.419963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.420011 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.420029 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.420052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.420077 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.523769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.523861 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.523893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.523925 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.523951 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.626434 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.626474 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.626485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.626499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.626510 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.729205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.729238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.729246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.729260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.729269 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.831792 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.831909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.831934 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.831967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.831989 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.934648 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.934704 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.934722 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.934746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:26 crc kubenswrapper[4983]: I1001 09:10:26.934763 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:26Z","lastTransitionTime":"2025-10-01T09:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.037504 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.037579 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.037598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.037622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.037639 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.140095 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.140138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.140152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.140175 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.140188 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.243338 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.243416 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.243438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.243464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.243491 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.346404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.346450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.346468 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.346492 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.346509 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.452622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.452696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.452717 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.452744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.452774 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.555909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.556070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.556092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.556118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.556135 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.660221 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.660282 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.660305 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.660333 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.660355 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.713833 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.713911 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.713846 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:27 crc kubenswrapper[4983]: E1001 09:10:27.714052 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.714209 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:27 crc kubenswrapper[4983]: E1001 09:10:27.714253 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:27 crc kubenswrapper[4983]: E1001 09:10:27.714339 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:27 crc kubenswrapper[4983]: E1001 09:10:27.714488 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.763657 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.763735 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.763770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.763801 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.763879 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.867169 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.867243 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.867263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.867286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.867306 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.971262 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.971338 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.971358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.971424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:27 crc kubenswrapper[4983]: I1001 09:10:27.971452 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:27Z","lastTransitionTime":"2025-10-01T09:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.074146 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.074205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.074223 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.074250 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.074267 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.177779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.177888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.177909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.177935 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.177954 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.281153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.281217 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.281236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.281262 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.281279 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.383876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.383927 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.383944 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.383968 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.383985 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.486747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.486797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.486856 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.486881 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.486902 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.590081 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.590169 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.590194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.590224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.590250 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.692573 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.692673 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.692696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.692730 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.692751 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.795933 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.796001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.796025 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.796057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.796082 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.879229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.879276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.879294 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.879316 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.879332 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: E1001 09:10:28.902586 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.908446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.908525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.908550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.908580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.908609 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: E1001 09:10:28.929626 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.935355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.935423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.935441 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.935469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.935489 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: E1001 09:10:28.956508 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.962113 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.962217 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.962237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.962261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.962291 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:28 crc kubenswrapper[4983]: E1001 09:10:28.984135 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.988790 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.988947 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.988969 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.988994 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:28 crc kubenswrapper[4983]: I1001 09:10:28.989012 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:28Z","lastTransitionTime":"2025-10-01T09:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.009285 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.009629 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.011761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.011891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.011911 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.011935 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.011958 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.116068 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.116124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.116141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.116165 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.116184 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.219849 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.219918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.219936 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.219963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.219980 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.323101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.323171 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.323189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.323215 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.323234 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.426336 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.426404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.426421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.426446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.426465 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.529440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.529516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.529540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.529572 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.529594 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.632235 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.632294 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.632312 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.632337 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.632355 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.714400 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.714453 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.714500 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.714634 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.714838 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.715044 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.715126 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:29 crc kubenswrapper[4983]: E1001 09:10:29.715239 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.736196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.736249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.736265 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.736288 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.736307 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.839621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.839683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.839701 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.839728 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.839746 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.943281 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.943380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.943398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.943422 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:29 crc kubenswrapper[4983]: I1001 09:10:29.943440 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:29Z","lastTransitionTime":"2025-10-01T09:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.046153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.046202 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.046234 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.046437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.046459 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.149848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.149900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.149916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.149976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.149996 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.253037 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.253099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.253117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.253142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.253159 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.356342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.356413 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.356437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.356473 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.356496 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.459638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.460109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.460249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.460396 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.460541 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.563946 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.564038 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.564058 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.564082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.564099 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.667385 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.667471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.667490 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.667518 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.667534 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.771017 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.771076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.771103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.771133 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.771153 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.874431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.874496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.874515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.874543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.874561 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.977388 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.977500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.977516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.977554 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:30 crc kubenswrapper[4983]: I1001 09:10:30.977567 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:30Z","lastTransitionTime":"2025-10-01T09:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.080386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.080420 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.080430 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.080446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.080458 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.183680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.183764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.183843 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.183877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.183902 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.286599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.286671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.286688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.286724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.286742 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.389497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.389565 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.389583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.389607 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.389624 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.492710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.492779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.492802 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.492875 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.492900 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.596510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.596591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.596614 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.596643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.596663 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.701922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.702000 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.702022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.702049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.702074 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.713613 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.713713 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.713767 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.713710 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:31 crc kubenswrapper[4983]: E1001 09:10:31.713965 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:31 crc kubenswrapper[4983]: E1001 09:10:31.714118 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:31 crc kubenswrapper[4983]: E1001 09:10:31.714273 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:31 crc kubenswrapper[4983]: E1001 09:10:31.714407 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.805446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.805534 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.805564 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.805600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.805624 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.908205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.908300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.908321 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.908343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:31 crc kubenswrapper[4983]: I1001 09:10:31.908397 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:31Z","lastTransitionTime":"2025-10-01T09:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.011183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.011249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.011266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.011290 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.011310 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.113845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.113889 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.113905 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.113929 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.113946 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.216714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.216773 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.216789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.216843 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.216869 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.320225 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.320300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.320318 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.320341 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.320361 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.423662 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.423708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.423725 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.423746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.423763 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.533231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.533289 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.533308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.533331 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.533350 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.636073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.636324 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.636373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.636407 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.636430 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.736128 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.738685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.738724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.738738 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.738756 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.738772 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.756182 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.774223 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.789998 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.814611 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.835250 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.840684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.840763 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.840789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.840858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.840887 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.854877 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.872114 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.897921 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.933060 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.943403 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.943439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.943451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.943465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.943476 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:32Z","lastTransitionTime":"2025-10-01T09:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.950650 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.968893 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.985748 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:32 crc kubenswrapper[4983]: I1001 09:10:32.999411 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.013883 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.028221 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.046201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.046253 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.046270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.046297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.046317 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.050221 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.062953 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.148894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.149442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.149577 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.149699 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.149789 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.253266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.253677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.253967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.254228 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.254416 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.357914 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.357974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.357986 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.358005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.358016 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.461053 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.461104 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.461121 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.461143 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.461161 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.564983 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.565044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.565064 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.565087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.565105 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.667319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.667403 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.667418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.667433 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.667472 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.714319 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.714372 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.714440 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:33 crc kubenswrapper[4983]: E1001 09:10:33.714440 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.714322 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:33 crc kubenswrapper[4983]: E1001 09:10:33.714501 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:33 crc kubenswrapper[4983]: E1001 09:10:33.714544 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:33 crc kubenswrapper[4983]: E1001 09:10:33.714588 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.769552 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.769605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.769620 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.769638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.769657 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.872198 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.872263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.872274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.872291 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.872303 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.974797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.974872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.974883 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.974900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:33 crc kubenswrapper[4983]: I1001 09:10:33.974911 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:33Z","lastTransitionTime":"2025-10-01T09:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.077695 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.077768 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.077787 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.077843 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.077862 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.181202 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.181258 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.181277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.181301 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.181318 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.284438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.284501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.284521 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.284547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.284565 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.387457 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.387496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.387504 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.387520 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.387529 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.490190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.490251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.490271 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.490297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.490314 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.593105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.593143 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.593154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.593185 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.593197 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.695401 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.695441 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.695452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.695475 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.695487 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.798283 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.798362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.798387 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.798421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.798445 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.902058 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.902128 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.902146 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.902170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:34 crc kubenswrapper[4983]: I1001 09:10:34.902188 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:34Z","lastTransitionTime":"2025-10-01T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.003764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.003793 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.003801 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.003826 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.003835 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.105928 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.105962 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.105974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.105989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.106001 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.208390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.208430 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.208441 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.208458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.208469 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.312127 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.312196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.312223 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.312257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.312280 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.415294 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.415390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.415443 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.415484 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.415594 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.517845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.517885 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.517893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.517909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.517918 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.621196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.621259 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.621284 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.621330 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.621365 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.714429 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.714536 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.714630 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:35 crc kubenswrapper[4983]: E1001 09:10:35.714726 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.714751 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:35 crc kubenswrapper[4983]: E1001 09:10:35.714899 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:35 crc kubenswrapper[4983]: E1001 09:10:35.715029 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:35 crc kubenswrapper[4983]: E1001 09:10:35.715146 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.715643 4983 scope.go:117] "RemoveContainer" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" Oct 01 09:10:35 crc kubenswrapper[4983]: E1001 09:10:35.715885 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.723439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.723477 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.723487 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.723502 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.723514 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.826882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.826950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.826969 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.826995 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.827015 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.929124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.929201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.929216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.929240 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:35 crc kubenswrapper[4983]: I1001 09:10:35.929258 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:35Z","lastTransitionTime":"2025-10-01T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.032170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.032227 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.032237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.032255 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.032268 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.134974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.135025 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.135033 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.135049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.135058 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.237106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.237181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.237200 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.237238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.237273 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.339015 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.339054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.339062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.339076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.339085 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.441050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.441139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.441151 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.441167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.441178 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.543469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.543503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.543514 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.543529 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.543541 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.646404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.646463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.646480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.646508 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.646524 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.749220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.749291 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.749315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.749345 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.749366 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.851991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.852073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.852099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.852127 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.852149 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.954865 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.954922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.954940 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.954963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:36 crc kubenswrapper[4983]: I1001 09:10:36.954980 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:36Z","lastTransitionTime":"2025-10-01T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.057653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.057715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.057733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.057755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.057772 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.159800 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.159850 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.159860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.159874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.159882 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.262509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.262550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.262561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.262574 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.262583 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.364869 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.364920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.364936 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.364951 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.364959 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.467464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.467519 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.467536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.467559 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.467577 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.570272 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.570344 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.570367 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.570396 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.570419 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.672413 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.672495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.672517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.672542 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.672560 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.713835 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.713879 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.713879 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.713838 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:37 crc kubenswrapper[4983]: E1001 09:10:37.714029 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:37 crc kubenswrapper[4983]: E1001 09:10:37.714101 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:37 crc kubenswrapper[4983]: E1001 09:10:37.714195 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:37 crc kubenswrapper[4983]: E1001 09:10:37.714272 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.775007 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.775061 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.775073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.775092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.775105 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.877631 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.877680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.877693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.877709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.877720 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.979773 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.979836 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.979848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.979863 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:37 crc kubenswrapper[4983]: I1001 09:10:37.979876 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:37Z","lastTransitionTime":"2025-10-01T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.082526 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.082569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.082581 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.082598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.082607 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.182901 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:38 crc kubenswrapper[4983]: E1001 09:10:38.183035 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:38 crc kubenswrapper[4983]: E1001 09:10:38.183087 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:10.183071306 +0000 UTC m=+98.172300103 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.184120 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.184149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.184158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.184171 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.184180 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.286392 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.286885 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.286894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.286906 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.286917 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.388964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.389008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.389017 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.389031 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.389040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.490926 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.490977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.490989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.491006 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.491016 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.593585 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.593634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.593647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.593666 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.593678 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.695944 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.695994 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.696007 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.696027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.696039 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.799040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.799111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.799130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.799158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.799194 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.904994 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.905486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.905583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.905680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:38 crc kubenswrapper[4983]: I1001 09:10:38.905777 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:38Z","lastTransitionTime":"2025-10-01T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.009472 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.010015 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.010118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.010182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.010245 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.112348 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.112416 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.112434 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.112458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.112475 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.215455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.215893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.215924 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.215939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.215948 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.274711 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.274763 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.274775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.274793 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.274822 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.287926 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.291967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.292062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.292074 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.292114 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.292129 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.304982 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.308967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.309022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.309039 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.309063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.309078 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.323326 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.327549 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.327591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.327603 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.327621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.327634 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.346404 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.351959 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.351998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.352035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.352054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.352067 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.365310 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.365470 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.367395 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.367457 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.367479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.367509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.367532 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.470675 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.470732 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.470745 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.470764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.470780 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.574208 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.574249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.574258 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.574272 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.574281 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.678166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.678233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.678254 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.678282 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.678301 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.714478 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.714546 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.714587 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.714490 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.714652 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.714851 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.714906 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:39 crc kubenswrapper[4983]: E1001 09:10:39.714997 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.782080 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.782136 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.782154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.782178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.782196 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.888799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.888876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.888893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.888916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.888933 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.992563 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.992622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.992646 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.992676 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:39 crc kubenswrapper[4983]: I1001 09:10:39.992710 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:39Z","lastTransitionTime":"2025-10-01T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.096215 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.096256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.096268 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.096284 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.096297 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.199047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.199113 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.199123 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.199136 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.199144 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.301373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.301415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.301428 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.301445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.301457 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.403100 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.403145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.403157 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.403180 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.403192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.505257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.505334 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.505346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.505365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.505376 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.607609 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.607876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.607991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.608082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.608165 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.709880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.709917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.709928 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.709943 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.709953 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.811957 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.811999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.812010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.812027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.812040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.914552 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.914581 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.914590 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.914602 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:40 crc kubenswrapper[4983]: I1001 09:10:40.914610 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:40Z","lastTransitionTime":"2025-10-01T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.017439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.017474 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.017483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.017497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.017507 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.119865 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.119901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.119912 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.119925 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.119934 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.125608 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/0.log" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.125659 4983 generic.go:334] "Generic (PLEG): container finished" podID="eb6d10ad-935a-496d-8c25-32e024026cd9" containerID="b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be" exitCode=1 Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.125687 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerDied","Data":"b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.126070 4983 scope.go:117] "RemoveContainer" containerID="b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.141350 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.154348 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.163667 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.175203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.196725 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.207616 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.221600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.221647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.221661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.221677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.221688 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.222053 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.235311 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.248407 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.260855 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.270629 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.283891 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.296133 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.307516 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.317958 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.323686 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.323854 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.323917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.323977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.324051 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.333077 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.352655 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.364549 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.425742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.425989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.426054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.426125 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.426194 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.528257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.528294 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.528304 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.528319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.528331 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.630776 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.630858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.630874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.630896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.630911 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.714610 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.714639 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:41 crc kubenswrapper[4983]: E1001 09:10:41.714747 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:41 crc kubenswrapper[4983]: E1001 09:10:41.714875 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.714946 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:41 crc kubenswrapper[4983]: E1001 09:10:41.714993 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.715036 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:41 crc kubenswrapper[4983]: E1001 09:10:41.715078 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.734245 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.734326 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.734351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.734388 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.734410 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.837547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.837604 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.837616 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.837634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.837646 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.939698 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.939743 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.939754 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.939771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:41 crc kubenswrapper[4983]: I1001 09:10:41.939783 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:41Z","lastTransitionTime":"2025-10-01T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.041719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.041759 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.041768 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.041782 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.041795 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.130699 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/0.log" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.130781 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerStarted","Data":"4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.144242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.144311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.144335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.144369 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.144394 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.151606 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.164053 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.175998 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.188261 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.199075 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.217319 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.232015 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.249050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.249099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.249111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.249125 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.249135 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.265887 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.294402 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.312347 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.331278 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.346686 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.351451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.351493 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.351507 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.351527 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.351541 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.365091 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.378120 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.389840 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.400953 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.414336 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.434072 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.453682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.453730 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.453743 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.453761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.453776 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.556465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.556511 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.556521 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.556540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.556623 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.658916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.658975 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.658988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.659004 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.659015 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.728226 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.741471 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.755576 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.760956 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.760998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.761010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.761028 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.761040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.766738 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.778488 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.791550 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.810354 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.826109 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.837901 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.848044 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.856471 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.863979 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.864017 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.864026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.864040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.864049 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.871457 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.888683 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.904089 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.918508 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.929578 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.943943 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.962947 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.966557 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.966595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.966603 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.966618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:42 crc kubenswrapper[4983]: I1001 09:10:42.966627 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:42Z","lastTransitionTime":"2025-10-01T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.069408 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.069451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.069462 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.069478 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.069489 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.170820 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.170850 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.170858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.170877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.170885 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.272831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.272867 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.272880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.272894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.272904 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.376383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.376432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.376450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.376473 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.376490 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.478843 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.478891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.478904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.478921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.478936 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.581884 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.581938 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.581954 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.581977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.581993 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.684483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.684522 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.684533 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.684548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.684560 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.713556 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:43 crc kubenswrapper[4983]: E1001 09:10:43.713661 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.713787 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:43 crc kubenswrapper[4983]: E1001 09:10:43.713871 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.713978 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:43 crc kubenswrapper[4983]: E1001 09:10:43.714025 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.714115 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:43 crc kubenswrapper[4983]: E1001 09:10:43.714160 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.786702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.786739 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.786748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.786765 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.786777 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.889432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.889471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.889482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.889501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.889514 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.991587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.991624 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.991635 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.991653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:43 crc kubenswrapper[4983]: I1001 09:10:43.991664 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:43Z","lastTransitionTime":"2025-10-01T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.094089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.094109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.094118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.094130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.094138 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.196661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.196699 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.196707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.196720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.196729 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.299627 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.299689 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.299713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.299738 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.299756 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.402304 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.402362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.402381 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.402411 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.402428 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.504346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.504708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.504886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.505041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.505177 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.608963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.609103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.609119 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.609140 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.609159 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.713486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.713517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.713525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.713539 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.713547 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.815439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.815487 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.815496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.815512 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.815522 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.917046 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.917087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.917099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.917116 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:44 crc kubenswrapper[4983]: I1001 09:10:44.917126 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:44Z","lastTransitionTime":"2025-10-01T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.019766 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.019849 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.019864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.019882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.019894 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.121956 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.121999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.122013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.122029 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.122040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.223882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.223937 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.223949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.223966 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.223978 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.326499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.326557 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.326568 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.326586 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.326599 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.429522 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.429574 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.429585 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.429600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.429610 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.532261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.532302 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.532315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.532328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.532337 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.634871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.634916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.634931 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.634949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.634960 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.714465 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.714510 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.714554 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:45 crc kubenswrapper[4983]: E1001 09:10:45.714578 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.714528 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:45 crc kubenswrapper[4983]: E1001 09:10:45.714694 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:45 crc kubenswrapper[4983]: E1001 09:10:45.714765 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:45 crc kubenswrapper[4983]: E1001 09:10:45.714836 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.737014 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.737052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.737063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.737081 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.737095 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.839996 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.840040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.840056 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.840080 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.840097 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.943268 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.943314 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.943334 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.943380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:45 crc kubenswrapper[4983]: I1001 09:10:45.943400 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:45Z","lastTransitionTime":"2025-10-01T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.045786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.045872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.045882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.045894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.045903 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.148021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.148050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.148059 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.148071 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.148079 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.250356 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.250440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.250459 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.250484 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.250534 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.353608 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.353697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.353716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.353770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.353792 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.457337 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.457404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.457419 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.457436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.457450 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.559953 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.560014 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.560031 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.560058 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.560076 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.662610 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.662663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.662680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.662700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.662714 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.764655 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.764719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.764737 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.764760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.764779 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.867941 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.868016 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.868040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.868066 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.868084 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.971110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.971166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.971184 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.971210 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:46 crc kubenswrapper[4983]: I1001 09:10:46.971226 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:46Z","lastTransitionTime":"2025-10-01T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.073896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.074030 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.074103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.074142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.074165 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.176917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.176991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.177016 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.177047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.177070 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.280332 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.280402 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.280421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.280445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.280462 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.383761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.383915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.383988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.384024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.384093 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.488049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.488108 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.488128 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.488156 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.488180 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.591682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.591757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.591777 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.591833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.591854 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.694782 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.694910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.695039 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.695060 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.695073 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.714397 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.714412 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.714558 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:47 crc kubenswrapper[4983]: E1001 09:10:47.714684 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.714797 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:47 crc kubenswrapper[4983]: E1001 09:10:47.714986 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:47 crc kubenswrapper[4983]: E1001 09:10:47.715679 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:47 crc kubenswrapper[4983]: E1001 09:10:47.715832 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.716554 4983 scope.go:117] "RemoveContainer" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.796636 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.796690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.796702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.796755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.796769 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.899108 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.899178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.899188 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.899241 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:47 crc kubenswrapper[4983]: I1001 09:10:47.899252 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:47Z","lastTransitionTime":"2025-10-01T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.001583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.001633 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.001649 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.001668 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.001682 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.104713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.104781 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.104801 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.104857 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.104875 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.151193 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/2.log" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.154624 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.155746 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.185797 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.207117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.207152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.207163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.207178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.207190 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.211134 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.236174 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.253864 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.266570 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.284797 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.295999 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.309864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.309899 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.309911 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.309926 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.309938 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.323549 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.340896 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.356986 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.370546 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.383904 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.401761 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.411894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.411935 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.411947 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.411964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.411975 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.414350 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.426285 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.443485 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.453194 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.468629 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.515601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.515635 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.515648 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.515666 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.515679 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.618357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.618481 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.618561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.618595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.618663 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.720485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.720530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.720541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.720559 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.720574 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.823015 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.823085 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.823093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.823106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.823114 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.926326 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.926377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.926393 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.926416 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:48 crc kubenswrapper[4983]: I1001 09:10:48.926432 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:48Z","lastTransitionTime":"2025-10-01T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.029688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.029739 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.029755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.029778 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.029794 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.132760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.132801 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.132894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.132917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.132929 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.160643 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/3.log" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.161311 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/2.log" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.165069 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" exitCode=1 Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.165110 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.165151 4983 scope.go:117] "RemoveContainer" containerID="ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.166292 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.166635 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.184862 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.198545 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.217518 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.234566 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.235485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.235546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.235572 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.235601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.235618 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.257253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.274975 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.292187 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.305479 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.324137 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.340443 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.340481 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.340491 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.340510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.340522 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.350105 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccb1fcc0f337dd2ad188f169d41990464d0f55c577491f1ca374b7352ea67359\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:20Z\\\",\\\"message\\\":\\\"informers/factory.go:160\\\\nI1001 09:10:20.677160 6647 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 09:10:20.677455 6647 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:10:20.677473 6647 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:10:20.677512 6647 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 09:10:20.677511 6647 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:10:20.677519 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 09:10:20.677551 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:10:20.677591 6647 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:10:20.677597 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 09:10:20.677617 6647 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:10:20.677626 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 09:10:20.677634 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:10:20.677659 6647 factory.go:656] Stopping watch factory\\\\nI1001 09:10:20.677676 6647 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:48Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:10:48.609499 7015 services_controller.go:434] Service openshift-console/downloads retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{downloads openshift-console bbc81ad7-5d87-40bf-82c5-a4db2311cff9 12322 0 2025-02-23 05:39:22 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[] map[operator.openshift.io/spec-hash:41d6e4f36bf41ab5be57dec2289f1f8807bbed4b0f642342f213a53bb3ff4d6d] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: console,component: downloads,},ClusterIP:10.217.4.213,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Clu\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.365449 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.380907 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.396640 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.409126 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.410764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.410829 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.410841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.410858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.410870 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.420598 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.429771 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.434840 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.434907 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.434919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.434935 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.434949 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.439409 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.450100 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.452451 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.455001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.455047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.455066 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.455094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.455112 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.468835 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.472199 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.476115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.476175 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.476193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.476217 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.476234 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.489414 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.493075 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.493109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.493118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.493138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.493148 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.507408 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:49Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.507557 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.509251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.509284 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.509296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.509315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.509328 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.612587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.612661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.612679 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.612707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.612726 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.714013 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.714051 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.714162 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.714010 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.714426 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.715105 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.714732 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:49 crc kubenswrapper[4983]: E1001 09:10:49.714676 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.715414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.715505 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.715579 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.715643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.715706 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.818754 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.818841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.818862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.818885 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.818902 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.921486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.921530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.921543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.921560 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:49 crc kubenswrapper[4983]: I1001 09:10:49.921572 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:49Z","lastTransitionTime":"2025-10-01T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.024027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.024090 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.024101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.024118 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.024130 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.127796 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.127879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.127896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.127920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.127939 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.170592 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/3.log" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.174669 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:10:50 crc kubenswrapper[4983]: E1001 09:10:50.174826 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.189051 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.203164 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.218131 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230019 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230756 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230857 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.230869 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.240493 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.271183 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.287124 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.303373 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.315717 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.328431 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.333440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.333479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.333491 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.333506 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.333519 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.343866 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.355180 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.369586 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.381374 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.394957 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.404448 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.418912 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.435356 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.435601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.435742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.435966 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.436003 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.439227 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:48Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:10:48.609499 7015 services_controller.go:434] Service openshift-console/downloads retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{downloads openshift-console bbc81ad7-5d87-40bf-82c5-a4db2311cff9 12322 0 2025-02-23 05:39:22 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[] map[operator.openshift.io/spec-hash:41d6e4f36bf41ab5be57dec2289f1f8807bbed4b0f642342f213a53bb3ff4d6d] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: console,component: downloads,},ClusterIP:10.217.4.213,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Clu\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:50Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.538410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.538455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.538465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.538485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.538495 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.640771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.640872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.640894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.640919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.640937 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.743445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.743485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.743498 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.743513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.743524 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.845937 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.845982 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.845992 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.846006 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.846019 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.948277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.948326 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.948337 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.948354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:50 crc kubenswrapper[4983]: I1001 09:10:50.948367 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:50Z","lastTransitionTime":"2025-10-01T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.051142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.051178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.051187 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.051200 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.051211 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.153477 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.153506 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.153514 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.153526 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.153535 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.262876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.262912 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.262921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.262934 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.262947 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.365440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.365486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.365543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.365585 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.365595 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.467749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.467867 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.467887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.467907 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.467922 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.570613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.570665 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.570678 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.570695 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.570707 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.673001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.673038 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.673048 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.673061 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.673071 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.713791 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.713861 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.713836 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:51 crc kubenswrapper[4983]: E1001 09:10:51.713961 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.713799 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:51 crc kubenswrapper[4983]: E1001 09:10:51.714067 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:51 crc kubenswrapper[4983]: E1001 09:10:51.714208 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:51 crc kubenswrapper[4983]: E1001 09:10:51.714299 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.775997 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.776034 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.776044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.776062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.776074 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.878534 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.878575 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.878584 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.878598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.878607 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.980783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.980860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.980882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.980902 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:51 crc kubenswrapper[4983]: I1001 09:10:51.980920 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:51Z","lastTransitionTime":"2025-10-01T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.083550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.083588 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.083600 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.083617 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.083628 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.186953 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.187042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.187065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.187093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.187116 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.289778 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.289849 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.289862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.289879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.289890 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.392788 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.393396 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.393411 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.393471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.393481 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.496101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.496149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.496160 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.496176 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.496187 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.599130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.599203 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.599224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.599252 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.599271 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.706007 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.706047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.706061 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.706077 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.706087 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.733953 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.749071 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77f570f7523db97ed2842914c2aa0d4f26ddb5630b1fbf377919bcfca19d9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5e177ec197e90bf62a5fa341511edcf3500ed0a131fe2cedd2540b8377eddd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.759319 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2jzdq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47732399-5753-4824-b164-525265f61461\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46e9254f6ccd84491a9deb17f13a986640741f541b796c5156eca902844664fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-md9b6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2jzdq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.768761 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvhnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.779341 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0c575660-d6c7-4d1c-aa74-c8d20088cbfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e1ef0d9065023f241d2f4666335773272be9c5633f9f2c2f416784de474d215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d7fd30e907683b9b6ced02df8244e0b19c67d25760ccefdc482d06bb20d978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef80b69f41a481273a9fc8fe7de00f177d2a7de743cbc050307fc5d00e7ba685\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5253db57391862992fde511cc64adf98ee7061ce3d6942677b0c05ed15da29\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.789500 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d0564b4-0a98-4fa8-8f8f-1b4f826c0b2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce74cb4985c1aa535bf831e4445d86e85b12490bed8f9f96b2e0aa9085c74750\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2132e33195d538eda450e7fc76592db96166cac90773016a5633eb0ec98aab09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31f787a460cd201584798ee9b6a5a2a35635eb0ddf21dc23deb765c6b4e7be8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac22cca2bec27cfda12a90fe89da09d7fa144cb9663012c75913f6d9290d08ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.801138 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ea7ba6d85d22955f9e8e846ce89ace09e3b3b29ccdb68297d276cca5133479\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.808020 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.808048 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.808057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.808070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.808080 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.811707 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d557036832d8a1b98652952e575fc05946ac93999c218b6cc63a805aa188936e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.821172 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rqrqw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9881be0c-c245-4096-bbd7-02081e1d73b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9787d6ba9d9796d82f666a670920fbbe2756132f78f4300ad542db4b7cc3223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zvd5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rqrqw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.833759 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p92ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6d10ad-935a-496d-8c25-32e024026cd9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:40Z\\\",\\\"message\\\":\\\"2025-10-01T09:09:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f\\\\n2025-10-01T09:09:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3ebb8e43-676d-4535-97a7-3acd599d239f to /host/opt/cni/bin/\\\\n2025-10-01T09:09:55Z [verbose] multus-daemon started\\\\n2025-10-01T09:09:55Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:10:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-879lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p92ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.843509 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf4a18e-a1a1-45e4-a577-e0442040b6c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7c8a6ef37e7e9870be778de6c6b1db59bd8a975abe7dd801da1b26886034304\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b5849ab5c10dfa235d8ac3447a707055e6e1433b52911c82c60b26ae95b606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:10:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4jlf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.865981 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3521a860-2b91-4e88-b412-35689a9d4e05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb58df10536cae2c262dee55f59f686a12770a08670812dfde5f8c1caed631ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de74091f33521e814addf9ebc94c6d3be7106052eac759035dfdebfacf23afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f11b99549855f85ab99128b813d606c33ab351559aae4f39d1f78a994ee5ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e32349bc91df70a34a40b347c5a2488e08819aa755af4164e6560d5ac671e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dffd2e4ac3b3721c530f24e1a252be3a39e1b9527f62dbe8556914fd83b32250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d784d46906558da921fb9f2195aea72ccb1d71d2fc91a4d550aceb9c40d8a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73b9a36340ac39db2b0467301bbe06a3ca60c08c113c96c27a301d3e236e27a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dde1204bfed8d0acb45a535d41c4d5998a88305a5c4b2b7221296f82a055b131\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.877567 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.891354 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.903260 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4affe98-5451-464f-af7e-6a43e5841e02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8243f97f5878ff95bc45846b4a1c41507e15497ad58693f04756f106d2df3dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5ckk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pd8zz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.911063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.911109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.911117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.911131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.911140 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:52Z","lastTransitionTime":"2025-10-01T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.916205 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf7bbd5-e570-44a5-b21a-636d54e2875a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26c38f91f935f6dce0ee8710c435ddaddf04d83479c45644971ab684db7b6f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5e6eed7e4865ea00be394ef80cb19d7266099cc45a3988c0e0fac9ea2200141\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5a545e4f9feb7e1a34e00722cf1a466f9a80e3c5b3927cf64f56961a480d3c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bddab54eabd123c1ce635103c17901693be6c7369fe262545a7bfc357de1bc06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66408fd472c83406e3fd7cf650d6191b11061103df3b53de2d184ffe9e010101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f2036f7248e29144e09725d3678715d953e60a608d5252dfb5141a39c71c6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec7fe7cdca51c0f826189e1773cf0d2009962ac602ff3919c1d11389ac417a0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vfm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mcdn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.943756 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e374b0f-544c-458e-9de6-5dd3149c3dd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:10:48Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:48Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:10:48.609499 7015 services_controller.go:434] Service openshift-console/downloads retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{downloads openshift-console bbc81ad7-5d87-40bf-82c5-a4db2311cff9 12322 0 2025-02-23 05:39:22 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[] map[operator.openshift.io/spec-hash:41d6e4f36bf41ab5be57dec2289f1f8807bbed4b0f642342f213a53bb3ff4d6d] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: console,component: downloads,},ClusterIP:10.217.4.213,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Clu\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:10:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lpzmm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7skk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:52 crc kubenswrapper[4983]: I1001 09:10:52.960375 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f702bbc7-ce29-4d21-a916-9c716b65c95d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3005553c6a8efd48a210e552fbe11a42a79f2bb881f529134e79bc027741b27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b83fc38f0a2d1622fddf72437d9b700662331b9c53c7408aacde82ccb1b96a70\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e472f3984101892df515e5e58faa5c91e55ce16415ccb72385942b717e1fcedf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed2de1b073eaec49768d2928ff20c1f31827fe2f435ef70f3661ebc38da1e00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21d3c1f6a55caa87fa01661c267f91e3c7488ad82dc864befbefa895b83b089f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:09:51Z\\\",\\\"message\\\":\\\"50435 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1001 09:09:51.250460 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1001 09:09:51.250477 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1001 09:09:51.250490 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1001 09:09:51.250507 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1001 09:09:51.250847 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\"\\\\nI1001 09:09:51.250867 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3142146180/tls.crt::/tmp/serving-cert-3142146180/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759309775\\\\\\\\\\\\\\\" (2025-10-01 09:09:35 +0000 UTC to 2025-10-31 09:09:36 +0000 UTC (now=2025-10-01 09:09:51.250835533 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251003 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759309776\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759309776\\\\\\\\\\\\\\\" (2025-10-01 08:09:36 +0000 UTC to 2026-10-01 08:09:36 +0000 UTC (now=2025-10-01 09:09:51.250979056 +0000 UTC))\\\\\\\"\\\\nI1001 09:09:51.251047 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1001 09:09:51.251073 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1001 09:09:51.251092 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1001 09:09:51.251743 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff9c688b2f527052ec513f84a70b4ec4d10979486b436e7162c001fc60b23ec\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:09:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad31a185394ac6259dfc677cdd47105faa4ba46a3361b6f3f987fe1b6714ed8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:09:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:52Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.013448 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.013514 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.013529 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.013544 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.013554 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.116485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.116533 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.116550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.116570 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.116587 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.219480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.219775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.219893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.219990 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.220079 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.323295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.323709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.323793 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.323901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.323977 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.426653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.426700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.426714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.426736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.426752 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.528900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.528940 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.528952 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.528967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.528978 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.632727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.632792 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.632834 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.632860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.632878 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.714413 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.714495 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:53 crc kubenswrapper[4983]: E1001 09:10:53.714629 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.714760 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:53 crc kubenswrapper[4983]: E1001 09:10:53.714930 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.714454 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:53 crc kubenswrapper[4983]: E1001 09:10:53.715440 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:53 crc kubenswrapper[4983]: E1001 09:10:53.715557 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.736135 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.736216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.736233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.736263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.736302 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.839985 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.840043 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.840064 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.840582 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.840634 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.946457 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.946517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.946535 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.946558 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:53 crc kubenswrapper[4983]: I1001 09:10:53.946576 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:53Z","lastTransitionTime":"2025-10-01T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.049158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.049605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.049749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.049951 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.050091 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.152555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.152605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.152618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.152634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.152645 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.255458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.255503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.255515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.255533 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.255545 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.357224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.357287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.357296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.357310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.357320 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.459637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.460341 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.460463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.460564 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.460655 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.563628 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.563675 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.563690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.563709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.563727 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.666353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.666391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.666401 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.666415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.666424 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.768265 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.768348 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.768360 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.768375 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.768410 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.870428 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.870466 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.870475 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.870488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.870497 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.972725 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.972757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.972767 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.972782 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:54 crc kubenswrapper[4983]: I1001 09:10:54.972792 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:54Z","lastTransitionTime":"2025-10-01T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.074638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.074677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.074688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.074716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.074727 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.177720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.177750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.177758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.177770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.177779 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.280301 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.280371 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.280390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.280416 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.280432 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.383881 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.383928 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.383939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.383960 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.383973 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.464379 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.464663 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.464630148 +0000 UTC m=+147.453858965 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.486950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.486981 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.486989 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.487002 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.487010 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.565427 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.565491 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.565524 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.565548 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565560 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565645 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.565627638 +0000 UTC m=+147.554856435 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565655 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565670 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565682 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565713 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.56570442 +0000 UTC m=+147.554933217 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565720 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565794 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565929 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565955 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.565876 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.565838473 +0000 UTC m=+147.555067310 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.566102 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.566090329 +0000 UTC m=+147.555319126 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.590336 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.590426 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.590445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.590468 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.590514 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.693694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.693747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.693765 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.693789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.693837 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.714287 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.714577 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.714313 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.714347 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.714775 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.714950 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.715073 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:55 crc kubenswrapper[4983]: E1001 09:10:55.715107 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.797358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.797442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.797467 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.797495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.797516 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.900026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.900111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.900129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.900155 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:55 crc kubenswrapper[4983]: I1001 09:10:55.900172 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:55Z","lastTransitionTime":"2025-10-01T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.002273 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.002346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.002370 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.002398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.002415 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.106292 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.106357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.106374 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.106398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.106415 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.209513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.209592 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.209622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.209654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.209678 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.312503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.312554 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.312565 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.312583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.312613 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.416437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.416496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.416513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.416537 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.416554 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.518590 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.518662 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.518682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.518710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.518731 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.621224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.621332 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.621384 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.621409 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.621426 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.725927 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.726221 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.726233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.726250 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.726261 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.829340 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.829400 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.829417 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.829440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.829457 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.932897 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.933606 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.933645 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.933671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:56 crc kubenswrapper[4983]: I1001 09:10:56.933690 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:56Z","lastTransitionTime":"2025-10-01T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.036864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.036926 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.036944 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.036970 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.036987 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.139707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.139748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.139786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.139814 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.139824 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.242584 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.242632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.242643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.242661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.242672 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.345401 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.345437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.345449 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.345465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.345477 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.448898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.448965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.448990 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.449019 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.449039 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.551646 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.551721 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.551744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.551776 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.551835 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.654505 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.654582 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.654608 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.654634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.654655 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.713840 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.713897 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.713948 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.713869 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:57 crc kubenswrapper[4983]: E1001 09:10:57.714068 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:57 crc kubenswrapper[4983]: E1001 09:10:57.714159 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:57 crc kubenswrapper[4983]: E1001 09:10:57.714232 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:57 crc kubenswrapper[4983]: E1001 09:10:57.714335 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.758065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.758110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.758127 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.758145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.758159 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.861246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.861313 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.861335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.861368 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.861391 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.964747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.964862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.964888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.964914 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:57 crc kubenswrapper[4983]: I1001 09:10:57.964931 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:57Z","lastTransitionTime":"2025-10-01T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.067170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.067211 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.067222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.067237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.067247 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.170027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.170083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.170093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.170116 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.170129 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.273965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.274057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.274071 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.274093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.274104 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.377882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.378033 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.378055 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.378094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.378113 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.481733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.481791 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.481839 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.481866 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.481878 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.584411 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.584467 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.584492 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.584513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.584525 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.687124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.687176 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.687189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.687205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.687217 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.790696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.790740 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.790749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.790770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.790786 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.893551 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.893604 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.893618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.893639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.893654 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.997048 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.997126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.997139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.997161 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:58 crc kubenswrapper[4983]: I1001 09:10:58.997174 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:58Z","lastTransitionTime":"2025-10-01T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.100655 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.100724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.100754 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.100786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.100854 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.203455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.203509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.203519 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.203538 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.203551 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.307186 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.307262 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.307318 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.307346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.307362 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.412487 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.412550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.412574 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.412603 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.412624 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.521699 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.521779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.521799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.521852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.521871 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.556758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.556828 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.556845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.556864 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.556880 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.577683 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.583140 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.583177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.583190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.583206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.583221 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.603309 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.609580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.609633 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.609652 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.609675 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.609694 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.627721 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.632818 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.632888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.632899 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.632922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.632934 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.654642 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.660291 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.660353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.660365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.660383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.660395 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.680244 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:10:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa9ee4eb-b268-4e5c-8a70-451c4e255816\\\",\\\"systemUUID\\\":\\\"77d2c1af-84b4-4e4f-813d-45ea11b761e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:10:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.680586 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.682928 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.683000 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.683019 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.683049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.683069 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.714397 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.714538 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.714613 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.714389 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.714718 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.714839 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.715032 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:10:59 crc kubenswrapper[4983]: E1001 09:10:59.715191 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.731663 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.786715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.786834 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.786853 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.786877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.786895 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.890700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.890745 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.890764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.890787 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.890829 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.994158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.994229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.994246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.994275 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:10:59 crc kubenswrapper[4983]: I1001 09:10:59.994295 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:10:59Z","lastTransitionTime":"2025-10-01T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.097799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.097901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.097919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.097944 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.097966 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.201715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.201778 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.201798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.201854 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.201874 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.305145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.305219 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.305239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.305264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.305280 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.408704 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.408775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.408794 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.408848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.408867 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.511940 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.512013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.512030 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.512436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.512495 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.618346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.618383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.618392 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.618410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.618421 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.720237 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.720297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.720309 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.720328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.720340 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.823786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.823873 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.823917 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.823949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.823974 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.926869 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.926918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.926939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.926965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:00 crc kubenswrapper[4983]: I1001 09:11:00.926984 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:00Z","lastTransitionTime":"2025-10-01T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.030758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.030842 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.030860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.030883 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.030902 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.134775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.135158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.135311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.135454 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.135573 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.241881 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.242089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.242112 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.242137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.242162 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.345210 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.345280 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.345302 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.345330 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.345352 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.447983 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.448047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.448062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.448084 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.448098 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.551078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.551143 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.551204 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.551266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.551294 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.654915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.654961 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.654972 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.654988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.655000 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.714321 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.714377 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.714419 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.714419 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:01 crc kubenswrapper[4983]: E1001 09:11:01.714824 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:01 crc kubenswrapper[4983]: E1001 09:11:01.714922 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:01 crc kubenswrapper[4983]: E1001 09:11:01.715027 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:01 crc kubenswrapper[4983]: E1001 09:11:01.715148 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.757903 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.757964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.757976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.757996 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.758009 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.861052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.861099 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.861117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.861141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.861188 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.964022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.964085 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.964103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.964129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:01 crc kubenswrapper[4983]: I1001 09:11:01.964148 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:01Z","lastTransitionTime":"2025-10-01T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.066669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.066720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.066736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.066758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.066776 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.169383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.169415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.169424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.169439 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.169448 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.272184 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.272349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.272373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.272401 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.272419 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.375192 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.375233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.375244 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.375260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.375271 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.478798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.478911 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.478940 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.478974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.478999 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.581624 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.581675 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.581692 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.581716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.581736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.684105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.684153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.684170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.684193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.684212 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.718351 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:11:02 crc kubenswrapper[4983]: E1001 09:11:02.718717 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.747346 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podStartSLOduration=70.747322262 podStartE2EDuration="1m10.747322262s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.746896562 +0000 UTC m=+90.736125389" watchObservedRunningTime="2025-10-01 09:11:02.747322262 +0000 UTC m=+90.736551099" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.787380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.787438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.787460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.787490 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.787514 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.807114 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mcdn7" podStartSLOduration=70.807089594 podStartE2EDuration="1m10.807089594s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.772339727 +0000 UTC m=+90.761568564" watchObservedRunningTime="2025-10-01 09:11:02.807089594 +0000 UTC m=+90.796318431" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.841903 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.841866642 podStartE2EDuration="1m11.841866642s" podCreationTimestamp="2025-10-01 09:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.838339104 +0000 UTC m=+90.827567981" watchObservedRunningTime="2025-10-01 09:11:02.841866642 +0000 UTC m=+90.831095489" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.891697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.891749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.891761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.891777 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.893267 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.924322 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.92429817 podStartE2EDuration="3.92429817s" podCreationTimestamp="2025-10-01 09:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.907744806 +0000 UTC m=+90.896973613" watchObservedRunningTime="2025-10-01 09:11:02.92429817 +0000 UTC m=+90.913526977" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.958575 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.958548974 podStartE2EDuration="1m7.958548974s" podCreationTimestamp="2025-10-01 09:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.957665802 +0000 UTC m=+90.946894619" watchObservedRunningTime="2025-10-01 09:11:02.958548974 +0000 UTC m=+90.947777791" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.988140 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2jzdq" podStartSLOduration=70.988119052 podStartE2EDuration="1m10.988119052s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:02.98804102 +0000 UTC m=+90.977269847" watchObservedRunningTime="2025-10-01 09:11:02.988119052 +0000 UTC m=+90.977347849" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.997122 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.997390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.997477 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.997555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:02 crc kubenswrapper[4983]: I1001 09:11:02.997643 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:02Z","lastTransitionTime":"2025-10-01T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.027263 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rqrqw" podStartSLOduration=72.027236878 podStartE2EDuration="1m12.027236878s" podCreationTimestamp="2025-10-01 09:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:03.013759442 +0000 UTC m=+91.002988249" watchObservedRunningTime="2025-10-01 09:11:03.027236878 +0000 UTC m=+91.016465685" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.028140 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-p92ll" podStartSLOduration=71.028130821 podStartE2EDuration="1m11.028130821s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:03.02689275 +0000 UTC m=+91.016121557" watchObservedRunningTime="2025-10-01 09:11:03.028130821 +0000 UTC m=+91.017359638" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.079271 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4jlf8" podStartSLOduration=71.079245616 podStartE2EDuration="1m11.079245616s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:03.048133629 +0000 UTC m=+91.037362457" watchObservedRunningTime="2025-10-01 09:11:03.079245616 +0000 UTC m=+91.068474423" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.099584 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=67.099565163 podStartE2EDuration="1m7.099565163s" podCreationTimestamp="2025-10-01 09:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:03.081178185 +0000 UTC m=+91.070407012" watchObservedRunningTime="2025-10-01 09:11:03.099565163 +0000 UTC m=+91.088793980" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100079 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.100073256 podStartE2EDuration="41.100073256s" podCreationTimestamp="2025-10-01 09:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:03.099979803 +0000 UTC m=+91.089208610" watchObservedRunningTime="2025-10-01 09:11:03.100073256 +0000 UTC m=+91.089302063" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100303 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100325 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.100341 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.203349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.203446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.203460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.203510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.203529 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.305913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.305976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.305988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.306008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.306055 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.408867 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.408925 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.408938 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.408958 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.408970 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.513088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.513138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.513158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.513181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.513196 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.616465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.616519 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.616531 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.616555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.616567 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.714553 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.714741 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:03 crc kubenswrapper[4983]: E1001 09:11:03.714987 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.715117 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:03 crc kubenswrapper[4983]: E1001 09:11:03.715179 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.715267 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:03 crc kubenswrapper[4983]: E1001 09:11:03.715356 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:03 crc kubenswrapper[4983]: E1001 09:11:03.715555 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.721047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.721094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.721108 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.721130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.721149 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.824677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.824736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.824753 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.824778 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.824797 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.928088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.928150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.928168 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.928195 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:03 crc kubenswrapper[4983]: I1001 09:11:03.928213 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:03Z","lastTransitionTime":"2025-10-01T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.031165 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.031311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.031342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.032129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.032161 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.134322 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.134354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.134363 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.134376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.134386 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.236370 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.236442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.236466 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.236497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.236514 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.339579 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.339620 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.339628 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.339641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.339651 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.442426 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.442455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.442464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.442479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.442489 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.545987 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.546047 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.546067 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.546093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.546110 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.650079 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.650154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.650177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.650209 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.650236 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.752386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.753026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.753062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.753089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.753106 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.855847 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.855918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.855939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.855961 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.855978 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.958497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.958554 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.958572 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.958599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:04 crc kubenswrapper[4983]: I1001 09:11:04.958616 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:04Z","lastTransitionTime":"2025-10-01T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.061396 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.061456 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.061469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.061485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.061500 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.164163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.164224 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.164242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.164268 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.164289 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.267551 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.267618 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.267641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.267672 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.267693 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.370300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.370398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.370424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.370452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.370471 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.473143 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.473194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.473210 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.473232 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.473249 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.577385 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.577482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.577499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.577526 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.577543 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.679893 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.679955 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.679977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.680005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.680022 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.713721 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:05 crc kubenswrapper[4983]: E1001 09:11:05.713838 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.713836 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.713879 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:05 crc kubenswrapper[4983]: E1001 09:11:05.713931 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.713919 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:05 crc kubenswrapper[4983]: E1001 09:11:05.714063 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:05 crc kubenswrapper[4983]: E1001 09:11:05.714245 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.782697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.782766 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.782785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.782851 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.782871 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.886210 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.886280 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.886302 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.886333 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.886356 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.989340 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.989393 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.989404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.989420 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:05 crc kubenswrapper[4983]: I1001 09:11:05.989736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:05Z","lastTransitionTime":"2025-10-01T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.092602 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.092647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.092661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.092680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.092695 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.195995 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.196070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.196092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.196136 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.196153 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.299353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.299412 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.299431 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.299460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.299478 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.407712 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.407755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.407767 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.407783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.407796 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.510266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.510299 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.510310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.510328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.510339 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.613119 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.613157 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.613167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.613185 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.613195 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.718009 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.718079 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.718102 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.718131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.718155 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.820933 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.820985 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.820997 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.821016 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.821027 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.924292 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.924359 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.924378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.924406 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:06 crc kubenswrapper[4983]: I1001 09:11:06.924426 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:06Z","lastTransitionTime":"2025-10-01T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.027290 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.027374 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.027410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.027444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.027465 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.130770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.130869 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.130895 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.130925 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.130949 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.233126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.233581 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.233613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.233634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.233649 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.336948 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.336998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.337010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.337029 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.337040 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.440027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.440090 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.440111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.440137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.440157 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.543333 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.543376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.543391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.543410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.543425 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.646353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.646419 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.646437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.646462 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.646483 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.714417 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.714451 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.714463 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.714573 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:07 crc kubenswrapper[4983]: E1001 09:11:07.715218 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:07 crc kubenswrapper[4983]: E1001 09:11:07.715078 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:07 crc kubenswrapper[4983]: E1001 09:11:07.714942 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:07 crc kubenswrapper[4983]: E1001 09:11:07.715414 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.749641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.749683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.749692 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.749705 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.749717 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.852443 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.852509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.852521 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.852539 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.852550 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.955196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.955231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.955239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.955252 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:07 crc kubenswrapper[4983]: I1001 09:11:07.955260 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:07Z","lastTransitionTime":"2025-10-01T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.058112 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.058158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.058173 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.058189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.058200 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.160192 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.160251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.160261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.160274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.160285 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.262872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.262910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.262919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.262933 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.262945 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.365414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.365478 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.365499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.365521 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.365539 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.468681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.468727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.468742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.468760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.468772 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.571317 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.571622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.571782 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.572001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.572134 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.674792 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.674847 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.674858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.674880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.674890 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.776787 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.776839 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.776849 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.776863 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.776873 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.879831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.879890 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.879907 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.879931 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.879949 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.982449 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.982510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.982527 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.982553 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:08 crc kubenswrapper[4983]: I1001 09:11:08.982575 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:08Z","lastTransitionTime":"2025-10-01T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.086090 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.086179 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.086198 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.086222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.086257 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.189017 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.189083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.189100 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.189125 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.189140 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.292762 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.292921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.292948 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.292981 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.293006 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.396469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.396530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.396549 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.396572 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.396589 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.500358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.500418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.500430 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.500471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.500486 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.603303 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.603363 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.603374 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.603394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.603406 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.705991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.706063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.706122 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.706147 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.706164 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.714354 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.714422 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.714472 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.714392 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:09 crc kubenswrapper[4983]: E1001 09:11:09.714546 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:09 crc kubenswrapper[4983]: E1001 09:11:09.714707 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:09 crc kubenswrapper[4983]: E1001 09:11:09.714903 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:09 crc kubenswrapper[4983]: E1001 09:11:09.715029 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.809733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.809794 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.809870 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.809898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.809916 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.895560 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.895633 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.895653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.895679 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.895698 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.920423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.920475 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.920485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.920499 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.920507 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:11:09Z","lastTransitionTime":"2025-10-01T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.949156 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b"] Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.950097 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.952391 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.953492 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.954381 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 01 09:11:09 crc kubenswrapper[4983]: I1001 09:11:09.954522 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.128349 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c500ef13-c11a-4675-bac7-6b7f9baa1110-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.128394 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.128433 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c500ef13-c11a-4675-bac7-6b7f9baa1110-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.128456 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c500ef13-c11a-4675-bac7-6b7f9baa1110-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.128482 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229028 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c500ef13-c11a-4675-bac7-6b7f9baa1110-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229130 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229172 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c500ef13-c11a-4675-bac7-6b7f9baa1110-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229226 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229266 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229333 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229421 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c500ef13-c11a-4675-bac7-6b7f9baa1110-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.229317 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c500ef13-c11a-4675-bac7-6b7f9baa1110-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: E1001 09:11:10.229515 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:11:10 crc kubenswrapper[4983]: E1001 09:11:10.229606 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs podName:7a2350dc-e1ed-4b83-8fbd-6b8321f34895 nodeName:}" failed. No retries permitted until 2025-10-01 09:12:14.229578418 +0000 UTC m=+162.218807255 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs") pod "network-metrics-daemon-pvhnl" (UID: "7a2350dc-e1ed-4b83-8fbd-6b8321f34895") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.231420 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c500ef13-c11a-4675-bac7-6b7f9baa1110-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.237114 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c500ef13-c11a-4675-bac7-6b7f9baa1110-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.258335 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c500ef13-c11a-4675-bac7-6b7f9baa1110-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mc24b\" (UID: \"c500ef13-c11a-4675-bac7-6b7f9baa1110\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:10 crc kubenswrapper[4983]: I1001 09:11:10.274961 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.249113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" event={"ID":"c500ef13-c11a-4675-bac7-6b7f9baa1110","Type":"ContainerStarted","Data":"d104deae9c534a087210281941a25936bf860400a40627daa3e616a5a29ce7f7"} Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.249203 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" event={"ID":"c500ef13-c11a-4675-bac7-6b7f9baa1110","Type":"ContainerStarted","Data":"cbaabf51e63923c24685445b887cc175bcb05ce3339ac29900ca9595fae4b80b"} Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.272488 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mc24b" podStartSLOduration=79.272456854 podStartE2EDuration="1m19.272456854s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:11.270193398 +0000 UTC m=+99.259422285" watchObservedRunningTime="2025-10-01 09:11:11.272456854 +0000 UTC m=+99.261685691" Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.713718 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.713793 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:11 crc kubenswrapper[4983]: E1001 09:11:11.713853 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.713939 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:11 crc kubenswrapper[4983]: E1001 09:11:11.714042 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:11 crc kubenswrapper[4983]: I1001 09:11:11.714114 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:11 crc kubenswrapper[4983]: E1001 09:11:11.714196 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:11 crc kubenswrapper[4983]: E1001 09:11:11.714309 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:13 crc kubenswrapper[4983]: I1001 09:11:13.714399 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:13 crc kubenswrapper[4983]: I1001 09:11:13.714749 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:13 crc kubenswrapper[4983]: I1001 09:11:13.714774 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:13 crc kubenswrapper[4983]: I1001 09:11:13.714794 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:13 crc kubenswrapper[4983]: I1001 09:11:13.715124 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:11:13 crc kubenswrapper[4983]: E1001 09:11:13.715553 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:11:13 crc kubenswrapper[4983]: E1001 09:11:13.715761 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:13 crc kubenswrapper[4983]: E1001 09:11:13.715798 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:13 crc kubenswrapper[4983]: E1001 09:11:13.715870 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:13 crc kubenswrapper[4983]: E1001 09:11:13.715955 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:15 crc kubenswrapper[4983]: I1001 09:11:15.713854 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:15 crc kubenswrapper[4983]: I1001 09:11:15.713968 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:15 crc kubenswrapper[4983]: E1001 09:11:15.714049 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:15 crc kubenswrapper[4983]: I1001 09:11:15.714064 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:15 crc kubenswrapper[4983]: I1001 09:11:15.714094 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:15 crc kubenswrapper[4983]: E1001 09:11:15.714304 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:15 crc kubenswrapper[4983]: E1001 09:11:15.714384 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:15 crc kubenswrapper[4983]: E1001 09:11:15.714567 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:17 crc kubenswrapper[4983]: I1001 09:11:17.713758 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:17 crc kubenswrapper[4983]: I1001 09:11:17.713758 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:17 crc kubenswrapper[4983]: E1001 09:11:17.714342 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:17 crc kubenswrapper[4983]: I1001 09:11:17.713859 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:17 crc kubenswrapper[4983]: I1001 09:11:17.713836 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:17 crc kubenswrapper[4983]: E1001 09:11:17.714470 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:17 crc kubenswrapper[4983]: E1001 09:11:17.714582 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:17 crc kubenswrapper[4983]: E1001 09:11:17.714723 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:19 crc kubenswrapper[4983]: I1001 09:11:19.715239 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:19 crc kubenswrapper[4983]: E1001 09:11:19.715578 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:19 crc kubenswrapper[4983]: I1001 09:11:19.716062 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:19 crc kubenswrapper[4983]: E1001 09:11:19.716207 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:19 crc kubenswrapper[4983]: I1001 09:11:19.716520 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:19 crc kubenswrapper[4983]: E1001 09:11:19.716632 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:19 crc kubenswrapper[4983]: I1001 09:11:19.716830 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:19 crc kubenswrapper[4983]: E1001 09:11:19.716923 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:21 crc kubenswrapper[4983]: I1001 09:11:21.713799 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:21 crc kubenswrapper[4983]: I1001 09:11:21.713845 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:21 crc kubenswrapper[4983]: I1001 09:11:21.713800 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:21 crc kubenswrapper[4983]: E1001 09:11:21.713939 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:21 crc kubenswrapper[4983]: E1001 09:11:21.714080 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:21 crc kubenswrapper[4983]: I1001 09:11:21.714266 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:21 crc kubenswrapper[4983]: E1001 09:11:21.714291 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:21 crc kubenswrapper[4983]: E1001 09:11:21.714511 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:23 crc kubenswrapper[4983]: I1001 09:11:23.714546 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:23 crc kubenswrapper[4983]: I1001 09:11:23.715800 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:23 crc kubenswrapper[4983]: I1001 09:11:23.715929 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:23 crc kubenswrapper[4983]: E1001 09:11:23.715987 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:23 crc kubenswrapper[4983]: I1001 09:11:23.716204 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:23 crc kubenswrapper[4983]: E1001 09:11:23.716210 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:23 crc kubenswrapper[4983]: E1001 09:11:23.716364 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:23 crc kubenswrapper[4983]: E1001 09:11:23.716445 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:25 crc kubenswrapper[4983]: I1001 09:11:25.714446 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:25 crc kubenswrapper[4983]: I1001 09:11:25.714478 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:25 crc kubenswrapper[4983]: I1001 09:11:25.714495 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:25 crc kubenswrapper[4983]: E1001 09:11:25.714645 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:25 crc kubenswrapper[4983]: E1001 09:11:25.714776 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:25 crc kubenswrapper[4983]: E1001 09:11:25.714938 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:25 crc kubenswrapper[4983]: I1001 09:11:25.715081 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:25 crc kubenswrapper[4983]: E1001 09:11:25.715221 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:25 crc kubenswrapper[4983]: I1001 09:11:25.716129 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:11:25 crc kubenswrapper[4983]: E1001 09:11:25.716361 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7skk_openshift-ovn-kubernetes(3e374b0f-544c-458e-9de6-5dd3149c3dd0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.301066 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/1.log" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.301617 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/0.log" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.301695 4983 generic.go:334] "Generic (PLEG): container finished" podID="eb6d10ad-935a-496d-8c25-32e024026cd9" containerID="4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01" exitCode=1 Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.301747 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerDied","Data":"4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01"} Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.301869 4983 scope.go:117] "RemoveContainer" containerID="b152505fe643ddc2c039611116686dae89dcd93c78a27dd568a80bc5ac7232be" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.302519 4983 scope.go:117] "RemoveContainer" containerID="4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01" Oct 01 09:11:27 crc kubenswrapper[4983]: E1001 09:11:27.302909 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-p92ll_openshift-multus(eb6d10ad-935a-496d-8c25-32e024026cd9)\"" pod="openshift-multus/multus-p92ll" podUID="eb6d10ad-935a-496d-8c25-32e024026cd9" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.713526 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.713648 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:27 crc kubenswrapper[4983]: E1001 09:11:27.713697 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.713869 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:27 crc kubenswrapper[4983]: E1001 09:11:27.714020 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:27 crc kubenswrapper[4983]: I1001 09:11:27.714048 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:27 crc kubenswrapper[4983]: E1001 09:11:27.714200 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:27 crc kubenswrapper[4983]: E1001 09:11:27.714424 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:28 crc kubenswrapper[4983]: I1001 09:11:28.307365 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/1.log" Oct 01 09:11:29 crc kubenswrapper[4983]: I1001 09:11:29.713864 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:29 crc kubenswrapper[4983]: I1001 09:11:29.713909 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:29 crc kubenswrapper[4983]: I1001 09:11:29.713886 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:29 crc kubenswrapper[4983]: I1001 09:11:29.713865 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:29 crc kubenswrapper[4983]: E1001 09:11:29.714014 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:29 crc kubenswrapper[4983]: E1001 09:11:29.714145 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:29 crc kubenswrapper[4983]: E1001 09:11:29.714261 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:29 crc kubenswrapper[4983]: E1001 09:11:29.714331 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:31 crc kubenswrapper[4983]: I1001 09:11:31.713555 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:31 crc kubenswrapper[4983]: I1001 09:11:31.713597 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:31 crc kubenswrapper[4983]: I1001 09:11:31.713579 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:31 crc kubenswrapper[4983]: I1001 09:11:31.713553 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:31 crc kubenswrapper[4983]: E1001 09:11:31.713710 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:31 crc kubenswrapper[4983]: E1001 09:11:31.713870 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:31 crc kubenswrapper[4983]: E1001 09:11:31.714028 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:31 crc kubenswrapper[4983]: E1001 09:11:31.714172 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:32 crc kubenswrapper[4983]: E1001 09:11:32.647941 4983 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 01 09:11:32 crc kubenswrapper[4983]: E1001 09:11:32.826142 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:11:33 crc kubenswrapper[4983]: I1001 09:11:33.714024 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:33 crc kubenswrapper[4983]: I1001 09:11:33.714099 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:33 crc kubenswrapper[4983]: I1001 09:11:33.714099 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:33 crc kubenswrapper[4983]: I1001 09:11:33.714241 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:33 crc kubenswrapper[4983]: E1001 09:11:33.714238 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:33 crc kubenswrapper[4983]: E1001 09:11:33.714430 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:33 crc kubenswrapper[4983]: E1001 09:11:33.714662 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:33 crc kubenswrapper[4983]: E1001 09:11:33.714906 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:35 crc kubenswrapper[4983]: I1001 09:11:35.713463 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:35 crc kubenswrapper[4983]: I1001 09:11:35.713535 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:35 crc kubenswrapper[4983]: I1001 09:11:35.713579 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:35 crc kubenswrapper[4983]: I1001 09:11:35.713594 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:35 crc kubenswrapper[4983]: E1001 09:11:35.713644 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:35 crc kubenswrapper[4983]: E1001 09:11:35.713793 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:35 crc kubenswrapper[4983]: E1001 09:11:35.713958 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:35 crc kubenswrapper[4983]: E1001 09:11:35.714058 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:37 crc kubenswrapper[4983]: I1001 09:11:37.713606 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:37 crc kubenswrapper[4983]: I1001 09:11:37.713628 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:37 crc kubenswrapper[4983]: I1001 09:11:37.713645 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:37 crc kubenswrapper[4983]: I1001 09:11:37.713854 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:37 crc kubenswrapper[4983]: E1001 09:11:37.714168 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:37 crc kubenswrapper[4983]: E1001 09:11:37.714351 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:37 crc kubenswrapper[4983]: E1001 09:11:37.714486 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:37 crc kubenswrapper[4983]: E1001 09:11:37.714613 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:37 crc kubenswrapper[4983]: I1001 09:11:37.715959 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:11:37 crc kubenswrapper[4983]: E1001 09:11:37.828445 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.361502 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/3.log" Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.364029 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerStarted","Data":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.364420 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.388187 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podStartSLOduration=106.388164323 podStartE2EDuration="1m46.388164323s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:38.386744817 +0000 UTC m=+126.375973614" watchObservedRunningTime="2025-10-01 09:11:38.388164323 +0000 UTC m=+126.377393110" Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.733690 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pvhnl"] Oct 01 09:11:38 crc kubenswrapper[4983]: I1001 09:11:38.733848 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:38 crc kubenswrapper[4983]: E1001 09:11:38.733986 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:39 crc kubenswrapper[4983]: I1001 09:11:39.714407 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:39 crc kubenswrapper[4983]: I1001 09:11:39.714454 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:39 crc kubenswrapper[4983]: I1001 09:11:39.714408 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:39 crc kubenswrapper[4983]: E1001 09:11:39.714547 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:39 crc kubenswrapper[4983]: E1001 09:11:39.714638 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:39 crc kubenswrapper[4983]: E1001 09:11:39.714776 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:40 crc kubenswrapper[4983]: I1001 09:11:40.714391 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:40 crc kubenswrapper[4983]: E1001 09:11:40.714587 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:41 crc kubenswrapper[4983]: I1001 09:11:41.713911 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:41 crc kubenswrapper[4983]: I1001 09:11:41.714017 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:41 crc kubenswrapper[4983]: E1001 09:11:41.714108 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:41 crc kubenswrapper[4983]: I1001 09:11:41.714027 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:41 crc kubenswrapper[4983]: E1001 09:11:41.714279 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:41 crc kubenswrapper[4983]: E1001 09:11:41.714410 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:42 crc kubenswrapper[4983]: I1001 09:11:42.713700 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:42 crc kubenswrapper[4983]: E1001 09:11:42.714218 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:42 crc kubenswrapper[4983]: I1001 09:11:42.714684 4983 scope.go:117] "RemoveContainer" containerID="4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01" Oct 01 09:11:42 crc kubenswrapper[4983]: E1001 09:11:42.829449 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:11:43 crc kubenswrapper[4983]: I1001 09:11:43.381708 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/1.log" Oct 01 09:11:43 crc kubenswrapper[4983]: I1001 09:11:43.381771 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerStarted","Data":"2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662"} Oct 01 09:11:43 crc kubenswrapper[4983]: I1001 09:11:43.714500 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:43 crc kubenswrapper[4983]: I1001 09:11:43.714565 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:43 crc kubenswrapper[4983]: I1001 09:11:43.714617 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:43 crc kubenswrapper[4983]: E1001 09:11:43.714709 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:43 crc kubenswrapper[4983]: E1001 09:11:43.714821 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:43 crc kubenswrapper[4983]: E1001 09:11:43.714961 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:44 crc kubenswrapper[4983]: I1001 09:11:44.714076 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:44 crc kubenswrapper[4983]: E1001 09:11:44.714353 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:45 crc kubenswrapper[4983]: I1001 09:11:45.714395 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:45 crc kubenswrapper[4983]: E1001 09:11:45.715141 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:45 crc kubenswrapper[4983]: I1001 09:11:45.714511 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:45 crc kubenswrapper[4983]: I1001 09:11:45.714463 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:45 crc kubenswrapper[4983]: E1001 09:11:45.715257 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:45 crc kubenswrapper[4983]: E1001 09:11:45.715364 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:46 crc kubenswrapper[4983]: I1001 09:11:46.714503 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:46 crc kubenswrapper[4983]: E1001 09:11:46.714703 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvhnl" podUID="7a2350dc-e1ed-4b83-8fbd-6b8321f34895" Oct 01 09:11:47 crc kubenswrapper[4983]: I1001 09:11:47.713500 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:47 crc kubenswrapper[4983]: I1001 09:11:47.713565 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:47 crc kubenswrapper[4983]: E1001 09:11:47.713703 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:11:47 crc kubenswrapper[4983]: I1001 09:11:47.713987 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:47 crc kubenswrapper[4983]: E1001 09:11:47.714143 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:11:47 crc kubenswrapper[4983]: E1001 09:11:47.714289 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:11:48 crc kubenswrapper[4983]: I1001 09:11:48.713953 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:11:48 crc kubenswrapper[4983]: I1001 09:11:48.717520 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 09:11:48 crc kubenswrapper[4983]: I1001 09:11:48.718159 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.713495 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.713578 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.713494 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.716703 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.717915 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.717924 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 09:11:49 crc kubenswrapper[4983]: I1001 09:11:49.718047 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.502258 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.541405 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bkjc6"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.542050 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.544879 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.545080 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.547101 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.547611 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.548042 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.553206 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.553522 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.553929 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.554957 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.555603 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.555632 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.556557 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.572030 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.572086 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.573403 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.573619 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.573885 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.573912 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.574334 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.574504 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.574657 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.574510 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.575226 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.574588 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.576306 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.576975 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.577309 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-tf547"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.577720 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.579721 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.579851 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.580641 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.584708 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-jjb8k"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.585115 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.586066 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.590084 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.608408 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.608933 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.609383 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.609580 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.614613 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.614866 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.624140 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.624622 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.625140 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.626842 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.631615 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655066 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47sb9\" (UniqueName: \"kubernetes.io/projected/c8199b4d-1870-489d-a48e-d8a0376f9090-kube-api-access-47sb9\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655098 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416da0ef-bf05-4226-aae9-692b0ccf210d-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655125 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-client\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655164 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655210 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-policies\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655234 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w266\" (UniqueName: \"kubernetes.io/projected/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-kube-api-access-2w266\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655248 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-encryption-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655264 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-oauth-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655278 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-serving-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655293 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-serving-cert\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655312 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655330 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgwfj\" (UniqueName: \"kubernetes.io/projected/46bfb1e9-ebaf-4e47-9983-71e6d6a01154-kube-api-access-tgwfj\") pod \"downloads-7954f5f757-jjb8k\" (UID: \"46bfb1e9-ebaf-4e47-9983-71e6d6a01154\") " pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655356 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655377 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-262xt\" (UniqueName: \"kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655394 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655412 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655430 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-serving-cert\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655450 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655472 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-console-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655487 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8jg\" (UniqueName: \"kubernetes.io/projected/af78a493-d342-4762-87e6-fdfdda346927-kube-api-access-lq8jg\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655517 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-audit\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655536 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-client\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655557 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-dir\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655579 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-trusted-ca-bundle\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655600 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-oauth-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655621 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655647 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655668 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mmdb\" (UniqueName: \"kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655692 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-image-import-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655715 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-service-ca\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655744 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-node-pullsecrets\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655766 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj9db\" (UniqueName: \"kubernetes.io/projected/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-kube-api-access-gj9db\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655788 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655887 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wrrk\" (UniqueName: \"kubernetes.io/projected/558325d7-8e74-4107-b574-ffb61eb4f445-kube-api-access-5wrrk\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655915 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/416da0ef-bf05-4226-aae9-692b0ccf210d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655939 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655964 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6599\" (UniqueName: \"kubernetes.io/projected/416da0ef-bf05-4226-aae9-692b0ccf210d-kube-api-access-t6599\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.655986 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.656007 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/558325d7-8e74-4107-b574-ffb61eb4f445-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.656029 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.656051 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-encryption-config\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.656074 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.656096 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-audit-dir\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.660560 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.660767 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.660845 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.660871 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661019 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661024 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661147 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661161 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661415 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.661560 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.664732 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.667029 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.668088 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.668310 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.668599 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.668864 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.670510 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-n26vk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.671074 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.671451 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.671738 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.673272 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.674742 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.676610 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.677139 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.695736 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gwghb"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.696622 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.697438 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.699614 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.707605 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.708052 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.708906 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.708990 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709083 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709200 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709327 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709500 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709638 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709773 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.709954 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710100 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710236 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710345 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710451 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710556 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710620 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710711 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710862 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.710938 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.711110 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.712113 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.712414 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.712862 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.714242 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.714293 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.714314 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.717612 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.717962 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.718079 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.720155 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.720209 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.720913 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.720953 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.721063 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.723014 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.723330 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.724043 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.724929 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.725174 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.726081 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.726478 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.728061 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.728645 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.729277 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jnpd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.729361 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.730970 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgxgp"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.731902 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.733161 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.733245 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.734000 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.738057 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.743558 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.745112 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.773951 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774264 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774768 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774826 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774845 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-node-pullsecrets\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774773 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-node-pullsecrets\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774931 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj9db\" (UniqueName: \"kubernetes.io/projected/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-kube-api-access-gj9db\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774952 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774978 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.774995 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775016 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gnc5\" (UniqueName: \"kubernetes.io/projected/f7c24919-2815-458f-af2e-11ca53444cc1-kube-api-access-5gnc5\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775040 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-service-ca\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775058 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775074 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c24919-2815-458f-af2e-11ca53444cc1-serving-cert\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775103 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775124 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wrrk\" (UniqueName: \"kubernetes.io/projected/558325d7-8e74-4107-b574-ffb61eb4f445-kube-api-access-5wrrk\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775152 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775172 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/416da0ef-bf05-4226-aae9-692b0ccf210d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775188 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775207 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6599\" (UniqueName: \"kubernetes.io/projected/416da0ef-bf05-4226-aae9-692b0ccf210d-kube-api-access-t6599\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775230 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775249 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775266 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775283 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/558325d7-8e74-4107-b574-ffb61eb4f445-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775300 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775315 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-encryption-config\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775332 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxcmk\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-kube-api-access-lxcmk\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775350 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775367 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-audit-dir\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775391 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47sb9\" (UniqueName: \"kubernetes.io/projected/c8199b4d-1870-489d-a48e-d8a0376f9090-kube-api-access-47sb9\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775413 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416da0ef-bf05-4226-aae9-692b0ccf210d-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775433 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-client\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775465 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775483 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbh98\" (UniqueName: \"kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775509 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775524 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775542 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w266\" (UniqueName: \"kubernetes.io/projected/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-kube-api-access-2w266\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775558 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775574 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-policies\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775592 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-oauth-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775607 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-serving-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775623 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-serving-cert\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775637 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-encryption-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775654 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775670 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775696 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775711 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-config\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775728 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775754 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775770 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afb7d248-4e64-47c3-a974-10a00e1f2fca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775787 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-config\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775818 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgwfj\" (UniqueName: \"kubernetes.io/projected/46bfb1e9-ebaf-4e47-9983-71e6d6a01154-kube-api-access-tgwfj\") pod \"downloads-7954f5f757-jjb8k\" (UID: \"46bfb1e9-ebaf-4e47-9983-71e6d6a01154\") " pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775836 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775853 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-262xt\" (UniqueName: \"kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775868 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775884 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv9cs\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775905 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775923 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-serving-cert\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775937 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775956 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775970 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-serving-cert\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.775988 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f745\" (UniqueName: \"kubernetes.io/projected/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-kube-api-access-8f745\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776008 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776022 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776037 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-console-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776086 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8jg\" (UniqueName: \"kubernetes.io/projected/af78a493-d342-4762-87e6-fdfdda346927-kube-api-access-lq8jg\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776081 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-service-ca\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776102 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776117 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776134 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776167 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776183 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-trusted-ca\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776201 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-audit\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776217 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-client\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776232 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-dir\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776247 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-trusted-ca-bundle\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776264 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776279 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/afb7d248-4e64-47c3-a974-10a00e1f2fca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776296 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776312 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-oauth-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776327 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776342 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mmdb\" (UniqueName: \"kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776359 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776375 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-image-import-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.776784 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.778015 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rpg24"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.778441 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.778830 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-image-import-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.778863 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.779329 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.779895 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.778836 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780286 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780367 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780456 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/416da0ef-bf05-4226-aae9-692b0ccf210d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780503 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780582 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.779107 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780727 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780798 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780873 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.779480 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780082 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.780114 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.781128 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.781331 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.782047 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.784094 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.784439 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.784717 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.784871 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.785169 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.785631 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.784929 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.786634 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-serving-cert\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.787694 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-trusted-ca-bundle\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.787861 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-oauth-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.788684 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c8199b4d-1870-489d-a48e-d8a0376f9090-audit-dir\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.788734 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.789649 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5wzgd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.789956 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.790411 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.790657 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.790946 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.791269 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.791529 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.792017 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-serving-cert\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.792462 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416da0ef-bf05-4226-aae9-692b0ccf210d-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.792540 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af78a493-d342-4762-87e6-fdfdda346927-console-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.793113 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-policies\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.793141 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.793748 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.794087 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.794791 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795013 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-audit-dir\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: E1001 09:11:50.795359 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.295343024 +0000 UTC m=+139.284571901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795901 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795918 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.796797 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.796911 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795915 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-audit\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795983 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795970 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af78a493-d342-4762-87e6-fdfdda346927-console-oauth-config\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.796027 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.796016 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.796035 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.797573 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.795969 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-etcd-client\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.798009 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-serving-ca\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.798283 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-encryption-config\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.798577 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-etcd-client\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.799969 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.802278 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.802675 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.803550 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.803801 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.803597 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.804271 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w5pt5"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.804332 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/558325d7-8e74-4107-b574-ffb61eb4f445-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.805018 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.805464 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-encryption-config\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.806668 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.807389 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fvslk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.807856 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.808719 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.809249 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8199b4d-1870-489d-a48e-d8a0376f9090-serving-cert\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.809375 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.812326 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.812385 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.812912 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bkjc6"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.813003 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.819860 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.819913 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.819929 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wrsrk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.820775 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.821659 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.827676 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.829114 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.833230 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gwghb"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.835793 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jnpd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.844015 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.844078 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.848422 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.851357 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.854878 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.856157 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.856792 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgxgp"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.858172 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jjb8k"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.859248 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.860325 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-n26vk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.861605 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-97dph"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.862370 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.862683 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.864920 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tf547"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.866821 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.868647 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.869950 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w5pt5"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.871187 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.872172 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fvslk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.873200 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.874219 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.875422 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.876417 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.876680 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:50 crc kubenswrapper[4983]: E1001 09:11:50.876781 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.37674449 +0000 UTC m=+139.365973287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.876931 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61c94092-2bfe-4a9a-ab80-c6268da4d42c-metrics-tls\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.876970 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-srv-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.876993 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-auth-proxy-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877059 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877084 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnbf\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-kube-api-access-hwnbf\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877102 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbh98\" (UniqueName: \"kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877129 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgf4r\" (UniqueName: \"kubernetes.io/projected/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-kube-api-access-pgf4r\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877147 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/027246b3-e611-429e-80c0-653dbde84d0a-tmpfs\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877190 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877283 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9v7f\" (UniqueName: \"kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877326 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkbd9\" (UniqueName: \"kubernetes.io/projected/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-kube-api-access-jkbd9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877377 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877417 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-config\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877441 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877461 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afb7d248-4e64-47c3-a974-10a00e1f2fca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877480 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-images\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877500 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-srv-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877521 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-profile-collector-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877544 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-config\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877562 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877583 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-metrics-tls\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877606 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6t2d\" (UniqueName: \"kubernetes.io/projected/50427e6c-1ddf-4141-9d2d-24ac5726c401-kube-api-access-t6t2d\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877636 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877658 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f745\" (UniqueName: \"kubernetes.io/projected/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-kube-api-access-8f745\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877680 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877704 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877726 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877747 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcm86\" (UniqueName: \"kubernetes.io/projected/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-kube-api-access-wcm86\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877778 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877821 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877839 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-trusted-ca\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877857 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877876 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877894 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cszq6\" (UniqueName: \"kubernetes.io/projected/9700b45d-c305-4dae-848f-c05699323db2-kube-api-access-cszq6\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877916 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqjwf\" (UniqueName: \"kubernetes.io/projected/01cf3320-f019-4bfb-9de1-826d0335ec36-kube-api-access-lqjwf\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877940 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhzrl\" (UniqueName: \"kubernetes.io/projected/e13daed8-fbbc-4c2e-93fc-411c659910e9-kube-api-access-nhzrl\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877956 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-config\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877973 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.877991 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878009 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878032 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878057 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c24919-2815-458f-af2e-11ca53444cc1-serving-cert\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878072 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72a01b33-7f81-470b-8c5c-1a2d12f40420-machine-approver-tls\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878098 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878115 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-webhook-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878141 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878159 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878180 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878196 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7k4h\" (UniqueName: \"kubernetes.io/projected/db154c97-53b6-4602-a6ab-4425c548f724-kube-api-access-b7k4h\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878214 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878231 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqccz\" (UniqueName: \"kubernetes.io/projected/05d0252d-cab0-454c-aa99-fe8f77e658bc-kube-api-access-zqccz\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878251 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878260 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-config\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878275 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878339 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxcmk\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-kube-api-access-lxcmk\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878376 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7f6v\" (UniqueName: \"kubernetes.io/projected/177c9b4a-8758-47be-9b2e-004485032310-kube-api-access-h7f6v\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878399 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8zpm\" (UniqueName: \"kubernetes.io/projected/2c6fc738-8526-4906-ad37-040910d43c72-kube-api-access-h8zpm\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878443 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878465 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj9db\" (UniqueName: \"kubernetes.io/projected/eba6b7fd-53c1-4d17-be7c-a35bb7e312d9-kube-api-access-gj9db\") pod \"openshift-apiserver-operator-796bbdcf4f-284pv\" (UID: \"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878468 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d0252d-cab0-454c-aa99-fe8f77e658bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878483 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878461 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878491 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878542 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afb7d248-4e64-47c3-a974-10a00e1f2fca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878554 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878580 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878715 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878764 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878853 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: E1001 09:11:50.878866 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.378857044 +0000 UTC m=+139.368085841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878889 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878914 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-serving-cert\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.878941 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879003 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879040 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv9cs\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879061 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879110 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879137 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-serving-cert\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879160 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c6fc738-8526-4906-ad37-040910d43c72-proxy-tls\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879187 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879591 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-client\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879626 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9700b45d-c305-4dae-848f-c05699323db2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879647 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-config\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879653 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879713 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879766 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879822 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-service-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879850 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/afb7d248-4e64-47c3-a974-10a00e1f2fca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879894 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c24919-2815-458f-af2e-11ca53444cc1-trusted-ca\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879953 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rpg24"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.879129 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.880431 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-service-ca-bundle\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.880464 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.880609 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.880756 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881407 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881471 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61c94092-2bfe-4a9a-ab80-c6268da4d42c-trusted-ca\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881491 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881510 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxbt5\" (UniqueName: \"kubernetes.io/projected/65fcd0cc-38b3-4951-8379-261a96372b6f-kube-api-access-bxbt5\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881573 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhtm\" (UniqueName: \"kubernetes.io/projected/027246b3-e611-429e-80c0-653dbde84d0a-kube-api-access-gxhtm\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881591 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881665 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881724 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gnc5\" (UniqueName: \"kubernetes.io/projected/f7c24919-2815-458f-af2e-11ca53444cc1-kube-api-access-5gnc5\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.881848 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575bm\" (UniqueName: \"kubernetes.io/projected/72a01b33-7f81-470b-8c5c-1a2d12f40420-kube-api-access-575bm\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.882534 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.882601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.883200 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-serving-cert\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.883234 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.883592 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.884116 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.884253 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.884309 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.884943 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.884952 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.885065 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.885223 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.885759 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c24919-2815-458f-af2e-11ca53444cc1-serving-cert\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.886063 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.886208 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wrsrk"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.887335 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-92fqg"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.887761 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/afb7d248-4e64-47c3-a974-10a00e1f2fca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.887925 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.888296 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.889905 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jzscr"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.890020 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.890765 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.890792 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jzscr"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.890856 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.891886 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-92fqg"] Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.896179 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wrrk\" (UniqueName: \"kubernetes.io/projected/558325d7-8e74-4107-b574-ffb61eb4f445-kube-api-access-5wrrk\") pod \"cluster-samples-operator-665b6dd947-6brq6\" (UID: \"558325d7-8e74-4107-b574-ffb61eb4f445\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.901797 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.903828 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.924897 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.943844 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.965015 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983247 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:50 crc kubenswrapper[4983]: E1001 09:11:50.983404 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.48338424 +0000 UTC m=+139.472613037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983452 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983473 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-serving-cert\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983494 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983525 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983542 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c6fc738-8526-4906-ad37-040910d43c72-proxy-tls\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983556 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983575 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-client\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983592 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9700b45d-c305-4dae-848f-c05699323db2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983623 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983642 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-service-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983660 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61c94092-2bfe-4a9a-ab80-c6268da4d42c-trusted-ca\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983680 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983700 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxbt5\" (UniqueName: \"kubernetes.io/projected/65fcd0cc-38b3-4951-8379-261a96372b6f-kube-api-access-bxbt5\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983729 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhtm\" (UniqueName: \"kubernetes.io/projected/027246b3-e611-429e-80c0-653dbde84d0a-kube-api-access-gxhtm\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983746 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983768 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575bm\" (UniqueName: \"kubernetes.io/projected/72a01b33-7f81-470b-8c5c-1a2d12f40420-kube-api-access-575bm\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983787 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61c94092-2bfe-4a9a-ab80-c6268da4d42c-metrics-tls\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983802 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-srv-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983857 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-auth-proxy-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983879 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983900 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnbf\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-kube-api-access-hwnbf\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983929 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgf4r\" (UniqueName: \"kubernetes.io/projected/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-kube-api-access-pgf4r\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983947 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/027246b3-e611-429e-80c0-653dbde84d0a-tmpfs\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983972 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9v7f\" (UniqueName: \"kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.983991 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkbd9\" (UniqueName: \"kubernetes.io/projected/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-kube-api-access-jkbd9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984010 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-images\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984025 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-srv-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984040 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-profile-collector-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984083 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-metrics-tls\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984106 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6t2d\" (UniqueName: \"kubernetes.io/projected/50427e6c-1ddf-4141-9d2d-24ac5726c401-kube-api-access-t6t2d\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984138 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984162 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcm86\" (UniqueName: \"kubernetes.io/projected/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-kube-api-access-wcm86\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984404 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984584 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/027246b3-e611-429e-80c0-653dbde84d0a-tmpfs\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: E1001 09:11:50.984822 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.484795106 +0000 UTC m=+139.474023903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.984193 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985085 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985107 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cszq6\" (UniqueName: \"kubernetes.io/projected/9700b45d-c305-4dae-848f-c05699323db2-kube-api-access-cszq6\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985361 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqjwf\" (UniqueName: \"kubernetes.io/projected/01cf3320-f019-4bfb-9de1-826d0335ec36-kube-api-access-lqjwf\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985395 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhzrl\" (UniqueName: \"kubernetes.io/projected/e13daed8-fbbc-4c2e-93fc-411c659910e9-kube-api-access-nhzrl\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985411 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-config\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985428 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985721 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.985745 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72a01b33-7f81-470b-8c5c-1a2d12f40420-machine-approver-tls\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.986785 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-webhook-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.986919 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.986945 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987013 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7k4h\" (UniqueName: \"kubernetes.io/projected/db154c97-53b6-4602-a6ab-4425c548f724-kube-api-access-b7k4h\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987125 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-metrics-tls\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987151 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987156 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9700b45d-c305-4dae-848f-c05699323db2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987242 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqccz\" (UniqueName: \"kubernetes.io/projected/05d0252d-cab0-454c-aa99-fe8f77e658bc-kube-api-access-zqccz\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987273 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7f6v\" (UniqueName: \"kubernetes.io/projected/177c9b4a-8758-47be-9b2e-004485032310-kube-api-access-h7f6v\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.987293 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8zpm\" (UniqueName: \"kubernetes.io/projected/2c6fc738-8526-4906-ad37-040910d43c72-kube-api-access-h8zpm\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.988030 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.988053 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d0252d-cab0-454c-aa99-fe8f77e658bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.988181 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:50 crc kubenswrapper[4983]: I1001 09:11:50.988421 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.005715 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.025301 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.030117 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.045568 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.065601 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.065941 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-serving-cert\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.078774 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6"] Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.081049 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-client\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.083959 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.086407 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-config\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.089202 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.089655 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.58963426 +0000 UTC m=+139.578863057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.089896 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.090271 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.590261836 +0000 UTC m=+139.579490703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.107169 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.115148 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.125061 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.134702 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50427e6c-1ddf-4141-9d2d-24ac5726c401-etcd-service-ca\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.145872 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.165030 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.180557 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv"] Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.183925 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.189326 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61c94092-2bfe-4a9a-ab80-c6268da4d42c-metrics-tls\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.190951 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.191100 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.691052477 +0000 UTC m=+139.680281274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.191221 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.191520 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.691511969 +0000 UTC m=+139.680740766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.210886 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.216128 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61c94092-2bfe-4a9a-ab80-c6268da4d42c-trusted-ca\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.224859 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.264901 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6599\" (UniqueName: \"kubernetes.io/projected/416da0ef-bf05-4226-aae9-692b0ccf210d-kube-api-access-t6599\") pod \"openshift-config-operator-7777fb866f-s4bh2\" (UID: \"416da0ef-bf05-4226-aae9-692b0ccf210d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.267854 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.277440 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-262xt\" (UniqueName: \"kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt\") pod \"controller-manager-879f6c89f-wnxqd\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.285694 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.292784 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.293386 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.793366688 +0000 UTC m=+139.782595485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.304826 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.326245 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.344619 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.364029 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.368709 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.383991 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.395426 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.395779 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.89576283 +0000 UTC m=+139.884991627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.403917 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.407529 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.409897 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" event={"ID":"558325d7-8e74-4107-b574-ffb61eb4f445","Type":"ContainerStarted","Data":"fe7d963fb5b269c551c691298446ff661432ed6e059834564c3dd0fee8a522de"} Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.409953 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" event={"ID":"558325d7-8e74-4107-b574-ffb61eb4f445","Type":"ContainerStarted","Data":"4f5c86e509f0609e4443174828a6933b3f3a4b7958f2330f05f0aaa7c3a27f0a"} Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.409966 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" event={"ID":"558325d7-8e74-4107-b574-ffb61eb4f445","Type":"ContainerStarted","Data":"779412ecef6bf9bc02f9b6633886311cae582e66d95a34dad727d5083752281c"} Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.411030 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" event={"ID":"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9","Type":"ContainerStarted","Data":"396168e97cf4d2acf4e7ad46011d36304b0b9713fc357bf40a3c90b8bd94ca5f"} Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.411058 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" event={"ID":"eba6b7fd-53c1-4d17-be7c-a35bb7e312d9","Type":"ContainerStarted","Data":"41d391df0ef52bb4b7529c5d3d83eb9ee14e56267fa65d4ad81ae0c56763cc78"} Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.420004 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2"] Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.424574 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 01 09:11:51 crc kubenswrapper[4983]: W1001 09:11:51.426031 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod416da0ef_bf05_4226_aae9_692b0ccf210d.slice/crio-46d7da66624614399b76fcbc4818e2e8195eca640e1cd9dbd2ee8b73a856ec35 WatchSource:0}: Error finding container 46d7da66624614399b76fcbc4818e2e8195eca640e1cd9dbd2ee8b73a856ec35: Status 404 returned error can't find the container with id 46d7da66624614399b76fcbc4818e2e8195eca640e1cd9dbd2ee8b73a856ec35 Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.443650 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.448759 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/72a01b33-7f81-470b-8c5c-1a2d12f40420-machine-approver-tls\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.464177 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.465601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-auth-proxy-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.484610 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.494511 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72a01b33-7f81-470b-8c5c-1a2d12f40420-config\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.496639 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.496782 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.996751255 +0000 UTC m=+139.985980072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.497047 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.497464 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:51.997453113 +0000 UTC m=+139.986681930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.503913 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.524567 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.543927 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.547233 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-srv-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.562990 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.567546 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.571553 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.576976 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01cf3320-f019-4bfb-9de1-826d0335ec36-profile-collector-cert\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.584907 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.598159 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.598748 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.098734217 +0000 UTC m=+140.087963014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.619256 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47sb9\" (UniqueName: \"kubernetes.io/projected/c8199b4d-1870-489d-a48e-d8a0376f9090-kube-api-access-47sb9\") pod \"apiserver-76f77b778f-bkjc6\" (UID: \"c8199b4d-1870-489d-a48e-d8a0376f9090\") " pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.652513 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w266\" (UniqueName: \"kubernetes.io/projected/3b1a6a6e-c4f8-4e5e-acbf-62220edee276-kube-api-access-2w266\") pod \"apiserver-7bbb656c7d-jclq9\" (UID: \"3b1a6a6e-c4f8-4e5e-acbf-62220edee276\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.664593 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.665194 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8jg\" (UniqueName: \"kubernetes.io/projected/af78a493-d342-4762-87e6-fdfdda346927-kube-api-access-lq8jg\") pod \"console-f9d7485db-tf547\" (UID: \"af78a493-d342-4762-87e6-fdfdda346927\") " pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.674786 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2c6fc738-8526-4906-ad37-040910d43c72-images\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.684133 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.700321 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.701600 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.200718418 +0000 UTC m=+140.189947215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.704310 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.715971 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.719553 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c6fc738-8526-4906-ad37-040910d43c72-proxy-tls\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:51 crc kubenswrapper[4983]: W1001 09:11:51.720606 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1f1eb5b_6daf_49db_a36d_32efc8e8f09a.slice/crio-4d7d0f2107916625ba2e4d91850ae896377dd8a7f4d1f5284cf1ff8930f338d2 WatchSource:0}: Error finding container 4d7d0f2107916625ba2e4d91850ae896377dd8a7f4d1f5284cf1ff8930f338d2: Status 404 returned error can't find the container with id 4d7d0f2107916625ba2e4d91850ae896377dd8a7f4d1f5284cf1ff8930f338d2 Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.737685 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mmdb\" (UniqueName: \"kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb\") pod \"route-controller-manager-6576b87f9c-ptgcj\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.745063 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.748218 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-apiservice-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.749956 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/027246b3-e611-429e-80c0-653dbde84d0a-webhook-cert\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.765782 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.773188 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.778545 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e13daed8-fbbc-4c2e-93fc-411c659910e9-srv-cert\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.786277 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.788956 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.801772 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.803278 4983 request.go:700] Waited for 1.006054674s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.803488 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.303468189 +0000 UTC m=+140.292696986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.803835 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.804433 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.304421994 +0000 UTC m=+140.293650791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.806676 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.824365 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.847998 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.864566 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.879987 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.902712 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgwfj\" (UniqueName: \"kubernetes.io/projected/46bfb1e9-ebaf-4e47-9983-71e6d6a01154-kube-api-access-tgwfj\") pod \"downloads-7954f5f757-jjb8k\" (UID: \"46bfb1e9-ebaf-4e47-9983-71e6d6a01154\") " pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.904622 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.907213 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.907423 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.40739544 +0000 UTC m=+140.396624237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.907530 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.907930 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.407916203 +0000 UTC m=+140.397145000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.920226 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.928689 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.942526 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bkjc6"] Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.944611 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.950149 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.957110 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d0252d-cab0-454c-aa99-fe8f77e658bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.964522 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.988779 4983 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.988856 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls podName:65fcd0cc-38b3-4951-8379-261a96372b6f nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.488837197 +0000 UTC m=+140.478065994 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-w5pt5" (UID: "65fcd0cc-38b3-4951-8379-261a96372b6f") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.989116 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.989840 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 01 09:11:51 crc kubenswrapper[4983]: I1001 09:11:51.990197 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990706 4983 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990742 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics podName:a199e85c-98a8-4563-81a6-196998a19005 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.490732436 +0000 UTC m=+140.479961233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics") pod "marketplace-operator-79b997595-xhsfh" (UID: "a199e85c-98a8-4563-81a6-196998a19005") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990757 4983 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990821 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key podName:db154c97-53b6-4602-a6ab-4425c548f724 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.490796698 +0000 UTC m=+140.480025495 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key") pod "service-ca-9c57cc56f-fvslk" (UID: "db154c97-53b6-4602-a6ab-4425c548f724") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990870 4983 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990889 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images podName:65fcd0cc-38b3-4951-8379-261a96372b6f nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.49088398 +0000 UTC m=+140.480112767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images") pod "machine-api-operator-5694c8668f-w5pt5" (UID: "65fcd0cc-38b3-4951-8379-261a96372b6f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990908 4983 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990917 4983 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990926 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config podName:65fcd0cc-38b3-4951-8379-261a96372b6f nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.490920841 +0000 UTC m=+140.480149738 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config") pod "machine-api-operator-5694c8668f-w5pt5" (UID: "65fcd0cc-38b3-4951-8379-261a96372b6f") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990946 4983 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990968 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls podName:177c9b4a-8758-47be-9b2e-004485032310 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.490949932 +0000 UTC m=+140.480178729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls") pod "dns-default-wrsrk" (UID: "177c9b4a-8758-47be-9b2e-004485032310") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990979 4983 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.990984 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca podName:a199e85c-98a8-4563-81a6-196998a19005 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.490977432 +0000 UTC m=+140.480206229 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca") pod "marketplace-operator-79b997595-xhsfh" (UID: "a199e85c-98a8-4563-81a6-196998a19005") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.991021 4983 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.991027 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle podName:db154c97-53b6-4602-a6ab-4425c548f724 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.491020443 +0000 UTC m=+140.480249230 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle") pod "service-ca-9c57cc56f-fvslk" (UID: "db154c97-53b6-4602-a6ab-4425c548f724") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:51 crc kubenswrapper[4983]: E1001 09:11:51.991296 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume podName:177c9b4a-8758-47be-9b2e-004485032310 nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.49128461 +0000 UTC m=+140.480513407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume") pod "dns-default-wrsrk" (UID: "177c9b4a-8758-47be-9b2e-004485032310") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.004949 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.009015 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.009519 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.509503355 +0000 UTC m=+140.498732152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.028403 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.047739 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.052492 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9"] Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.064623 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: W1001 09:11:52.066334 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1a6a6e_c4f8_4e5e_acbf_62220edee276.slice/crio-73b096f69d71dd52ad21b9294c0258cd0a45012b7fd55f6b212fea7735b9bcec WatchSource:0}: Error finding container 73b096f69d71dd52ad21b9294c0258cd0a45012b7fd55f6b212fea7735b9bcec: Status 404 returned error can't find the container with id 73b096f69d71dd52ad21b9294c0258cd0a45012b7fd55f6b212fea7735b9bcec Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.084238 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.104506 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.106889 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tf547"] Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.110784 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.111103 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.611090466 +0000 UTC m=+140.600319263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.124105 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.144441 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.164416 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.190259 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.203979 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.211590 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.211728 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.711709032 +0000 UTC m=+140.700937829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.212100 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.212431 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.712420131 +0000 UTC m=+140.701648928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.223905 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.243916 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.264629 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.284222 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.304151 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.313861 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.314008 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.813987551 +0000 UTC m=+140.803216348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.314127 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.314431 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.814423963 +0000 UTC m=+140.803652760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.323955 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.345159 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.359984 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jjb8k"] Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.364348 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.385268 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.407102 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.414946 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tf547" event={"ID":"af78a493-d342-4762-87e6-fdfdda346927","Type":"ContainerStarted","Data":"75cce9e59737a4f5058fe4e416a1f8da711a94ff19d4bccafde0cd718424d113"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.414992 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tf547" event={"ID":"af78a493-d342-4762-87e6-fdfdda346927","Type":"ContainerStarted","Data":"0fbedb7d537671e400e47994c850f9230b07d2e3084c92590557b68a853887d6"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.415202 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.415471 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.91544935 +0000 UTC m=+140.904678147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.415932 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" event={"ID":"d081f5a5-7101-403d-82f5-4395d1db959e","Type":"ContainerStarted","Data":"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.415969 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" event={"ID":"d081f5a5-7101-403d-82f5-4395d1db959e","Type":"ContainerStarted","Data":"d5eda2896009250a7ad4c1a1b6ab4d275f609ccdce6ac7aecc6b2c7e31914ff2"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.416132 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.416454 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.417349 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:52.917334298 +0000 UTC m=+140.906563095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.417579 4983 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ptgcj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.417612 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.417693 4983 generic.go:334] "Generic (PLEG): container finished" podID="416da0ef-bf05-4226-aae9-692b0ccf210d" containerID="0f0ac4ceb9c6e33698f21c5ced5dcef74b7c84b08ba2a1acad4fbdf2dbbfcbc2" exitCode=0 Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.417926 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" event={"ID":"416da0ef-bf05-4226-aae9-692b0ccf210d","Type":"ContainerDied","Data":"0f0ac4ceb9c6e33698f21c5ced5dcef74b7c84b08ba2a1acad4fbdf2dbbfcbc2"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.417949 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" event={"ID":"416da0ef-bf05-4226-aae9-692b0ccf210d","Type":"ContainerStarted","Data":"46d7da66624614399b76fcbc4818e2e8195eca640e1cd9dbd2ee8b73a856ec35"} Oct 01 09:11:52 crc kubenswrapper[4983]: W1001 09:11:52.419115 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46bfb1e9_ebaf_4e47_9983_71e6d6a01154.slice/crio-8c557674a1dadf2efc647eabb562cfd49386f9cbef3eb306e5288a0657413f90 WatchSource:0}: Error finding container 8c557674a1dadf2efc647eabb562cfd49386f9cbef3eb306e5288a0657413f90: Status 404 returned error can't find the container with id 8c557674a1dadf2efc647eabb562cfd49386f9cbef3eb306e5288a0657413f90 Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.419607 4983 generic.go:334] "Generic (PLEG): container finished" podID="3b1a6a6e-c4f8-4e5e-acbf-62220edee276" containerID="b8a0cee4939fa3b173bd93ef6d697cff4c92bb5c5daae45d01fc88961446478a" exitCode=0 Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.419665 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" event={"ID":"3b1a6a6e-c4f8-4e5e-acbf-62220edee276","Type":"ContainerDied","Data":"b8a0cee4939fa3b173bd93ef6d697cff4c92bb5c5daae45d01fc88961446478a"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.419687 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" event={"ID":"3b1a6a6e-c4f8-4e5e-acbf-62220edee276","Type":"ContainerStarted","Data":"73b096f69d71dd52ad21b9294c0258cd0a45012b7fd55f6b212fea7735b9bcec"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.421710 4983 generic.go:334] "Generic (PLEG): container finished" podID="c8199b4d-1870-489d-a48e-d8a0376f9090" containerID="d57cdeb4ef51221b646ceadd0fbcc7a00b79a6bfd36585da5aa25407e1e40c03" exitCode=0 Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.421987 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" event={"ID":"c8199b4d-1870-489d-a48e-d8a0376f9090","Type":"ContainerDied","Data":"d57cdeb4ef51221b646ceadd0fbcc7a00b79a6bfd36585da5aa25407e1e40c03"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.422030 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" event={"ID":"c8199b4d-1870-489d-a48e-d8a0376f9090","Type":"ContainerStarted","Data":"135d49b88177576f370a1792ac53d9473f4ca01ffb97bd54c90c998911b6bcf9"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.424904 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" event={"ID":"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a","Type":"ContainerStarted","Data":"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.424932 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" event={"ID":"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a","Type":"ContainerStarted","Data":"4d7d0f2107916625ba2e4d91850ae896377dd8a7f4d1f5284cf1ff8930f338d2"} Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.424959 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.447348 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.464466 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.485011 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.504753 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519266 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.519466 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.019438492 +0000 UTC m=+141.008667299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519537 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519671 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519761 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519800 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519893 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.519939 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.520021 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.520067 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.020055288 +0000 UTC m=+141.009284085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.520112 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.520147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.520193 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.520989 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-config\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.521167 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.521204 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/65fcd0cc-38b3-4951-8379-261a96372b6f-images\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.522747 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db154c97-53b6-4602-a6ab-4425c548f724-signing-cabundle\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.523657 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/65fcd0cc-38b3-4951-8379-261a96372b6f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.524176 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db154c97-53b6-4602-a6ab-4425c548f724-signing-key\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.524777 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.525519 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.544908 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.564506 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.584836 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.591771 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/177c9b4a-8758-47be-9b2e-004485032310-config-volume\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.604197 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.614907 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/177c9b4a-8758-47be-9b2e-004485032310-metrics-tls\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.620639 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.620968 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.120923551 +0000 UTC m=+141.110152348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.621213 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.621527 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.121516356 +0000 UTC m=+141.110745153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.664072 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.683997 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.704081 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.722545 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.722716 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.222695497 +0000 UTC m=+141.211924294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.722921 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.723359 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.223350674 +0000 UTC m=+141.212579461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.746091 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbh98\" (UniqueName: \"kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98\") pod \"oauth-openshift-558db77b4-z72rd\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.761561 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxcmk\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-kube-api-access-lxcmk\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.785289 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f745\" (UniqueName: \"kubernetes.io/projected/7f6005e9-43e4-4cf2-9428-fb48a2a26f4d-kube-api-access-8f745\") pod \"authentication-operator-69f744f599-gwghb\" (UID: \"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.800975 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.820259 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afb7d248-4e64-47c3-a974-10a00e1f2fca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-49krl\" (UID: \"afb7d248-4e64-47c3-a974-10a00e1f2fca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.823903 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.824320 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.324303849 +0000 UTC m=+141.313532646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.824351 4983 request.go:700] Waited for 1.942400437s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.839435 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gnc5\" (UniqueName: \"kubernetes.io/projected/f7c24919-2815-458f-af2e-11ca53444cc1-kube-api-access-5gnc5\") pod \"console-operator-58897d9998-n26vk\" (UID: \"f7c24919-2815-458f-af2e-11ca53444cc1\") " pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.858729 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv9cs\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.867870 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.884237 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.904678 4983 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.924682 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.924988 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:52 crc kubenswrapper[4983]: E1001 09:11:52.925411 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.425399408 +0000 UTC m=+141.414628205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.934116 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.944451 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.947974 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.957260 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.963479 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.969839 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 01 09:11:52 crc kubenswrapper[4983]: I1001 09:11:52.985231 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.027641 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.028004 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.527987824 +0000 UTC m=+141.517216611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.028294 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.028572 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.528565169 +0000 UTC m=+141.517793966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.030873 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.057387 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575bm\" (UniqueName: \"kubernetes.io/projected/72a01b33-7f81-470b-8c5c-1a2d12f40420-kube-api-access-575bm\") pod \"machine-approver-56656f9798-rgz6d\" (UID: \"72a01b33-7f81-470b-8c5c-1a2d12f40420\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.057638 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhtm\" (UniqueName: \"kubernetes.io/projected/027246b3-e611-429e-80c0-653dbde84d0a-kube-api-access-gxhtm\") pod \"packageserver-d55dfcdfc-tg26f\" (UID: \"027246b3-e611-429e-80c0-653dbde84d0a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.083845 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxbt5\" (UniqueName: \"kubernetes.io/projected/65fcd0cc-38b3-4951-8379-261a96372b6f-kube-api-access-bxbt5\") pod \"machine-api-operator-5694c8668f-w5pt5\" (UID: \"65fcd0cc-38b3-4951-8379-261a96372b6f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.100105 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9v7f\" (UniqueName: \"kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f\") pod \"marketplace-operator-79b997595-xhsfh\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.103507 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.114064 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.120245 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkbd9\" (UniqueName: \"kubernetes.io/projected/59073ee5-3101-46ac-88b8-ad59cdb2f5a8-kube-api-access-jkbd9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mjxft\" (UID: \"59073ee5-3101-46ac-88b8-ad59cdb2f5a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.129427 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.129740 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.62972552 +0000 UTC m=+141.618954317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.129827 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.130104 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.630097069 +0000 UTC m=+141.619325866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.145875 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcm86\" (UniqueName: \"kubernetes.io/projected/1cdabf47-3a38-4c24-b2e8-40ae2f062f08-kube-api-access-wcm86\") pod \"dns-operator-744455d44c-4jnpd\" (UID: \"1cdabf47-3a38-4c24-b2e8-40ae2f062f08\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.165661 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgf4r\" (UniqueName: \"kubernetes.io/projected/dd4036a5-2db8-4cec-98c1-7723eebcfdcd-kube-api-access-pgf4r\") pod \"package-server-manager-789f6589d5-b2j8m\" (UID: \"dd4036a5-2db8-4cec-98c1-7723eebcfdcd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.180937 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.185612 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.193426 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnbf\" (UniqueName: \"kubernetes.io/projected/61c94092-2bfe-4a9a-ab80-c6268da4d42c-kube-api-access-hwnbf\") pod \"ingress-operator-5b745b69d9-qppm8\" (UID: \"61c94092-2bfe-4a9a-ab80-c6268da4d42c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.197124 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.225305 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cszq6\" (UniqueName: \"kubernetes.io/projected/9700b45d-c305-4dae-848f-c05699323db2-kube-api-access-cszq6\") pod \"multus-admission-controller-857f4d67dd-fgxgp\" (UID: \"9700b45d-c305-4dae-848f-c05699323db2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.231272 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.231519 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.731503736 +0000 UTC m=+141.720732523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.231572 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.231977 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.731944927 +0000 UTC m=+141.721173724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.236559 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqjwf\" (UniqueName: \"kubernetes.io/projected/01cf3320-f019-4bfb-9de1-826d0335ec36-kube-api-access-lqjwf\") pod \"catalog-operator-68c6474976-4xzjs\" (UID: \"01cf3320-f019-4bfb-9de1-826d0335ec36\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.256960 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhzrl\" (UniqueName: \"kubernetes.io/projected/e13daed8-fbbc-4c2e-93fc-411c659910e9-kube-api-access-nhzrl\") pod \"olm-operator-6b444d44fb-5mdqr\" (UID: \"e13daed8-fbbc-4c2e-93fc-411c659910e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.284681 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7k4h\" (UniqueName: \"kubernetes.io/projected/db154c97-53b6-4602-a6ab-4425c548f724-kube-api-access-b7k4h\") pod \"service-ca-9c57cc56f-fvslk\" (UID: \"db154c97-53b6-4602-a6ab-4425c548f724\") " pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.301775 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqccz\" (UniqueName: \"kubernetes.io/projected/05d0252d-cab0-454c-aa99-fe8f77e658bc-kube-api-access-zqccz\") pod \"control-plane-machine-set-operator-78cbb6b69f-5svzw\" (UID: \"05d0252d-cab0-454c-aa99-fe8f77e658bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.313470 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8zpm\" (UniqueName: \"kubernetes.io/projected/2c6fc738-8526-4906-ad37-040910d43c72-kube-api-access-h8zpm\") pod \"machine-config-operator-74547568cd-hsczv\" (UID: \"2c6fc738-8526-4906-ad37-040910d43c72\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.324321 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.328291 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7f6v\" (UniqueName: \"kubernetes.io/projected/177c9b4a-8758-47be-9b2e-004485032310-kube-api-access-h7f6v\") pod \"dns-default-wrsrk\" (UID: \"177c9b4a-8758-47be-9b2e-004485032310\") " pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.335528 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.339295 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.339603 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.839588693 +0000 UTC m=+141.828817490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.358514 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.359608 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6t2d\" (UniqueName: \"kubernetes.io/projected/50427e6c-1ddf-4141-9d2d-24ac5726c401-kube-api-access-t6t2d\") pod \"etcd-operator-b45778765-rpg24\" (UID: \"50427e6c-1ddf-4141-9d2d-24ac5726c401\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.365499 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.377085 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.400566 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.420211 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.425530 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gwghb"] Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.431560 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441618 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6bc3891f-6cc8-445d-b852-6b34df2bf821-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441675 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwmbn\" (UniqueName: \"kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441694 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6deda0d-2483-4468-9959-7d7b47dd731d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441711 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/add58d5d-445b-47e0-9399-180a0c4571a8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441751 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-serving-cert\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441790 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2df46f36-de11-4b06-86d4-7ad0e6e58737-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441822 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcpmh\" (UniqueName: \"kubernetes.io/projected/2df46f36-de11-4b06-86d4-7ad0e6e58737-kube-api-access-kcpmh\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441839 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bc3891f-6cc8-445d-b852-6b34df2bf821-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441911 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqbk\" (UniqueName: \"kubernetes.io/projected/7330817e-3c60-4eaa-b546-76a64214a03f-kube-api-access-sgqbk\") pod \"migrator-59844c95c7-n5stn\" (UID: \"7330817e-3c60-4eaa-b546-76a64214a03f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441928 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6deda0d-2483-4468-9959-7d7b47dd731d-proxy-tls\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441946 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjn7n\" (UniqueName: \"kubernetes.io/projected/a6deda0d-2483-4468-9959-7d7b47dd731d-kube-api-access-xjn7n\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441979 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-default-certificate\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441995 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-stats-auth\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442030 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2df46f36-de11-4b06-86d4-7ad0e6e58737-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442053 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442071 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b287991-894d-4985-b0e0-74a46ae4a7d3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442104 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc3891f-6cc8-445d-b852-6b34df2bf821-config\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442120 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-config\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442145 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72kv\" (UniqueName: \"kubernetes.io/projected/11a442c1-aca4-467a-bbb7-5ccea25def0b-kube-api-access-r72kv\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442197 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442221 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-metrics-certs\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442245 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11a442c1-aca4-467a-bbb7-5ccea25def0b-service-ca-bundle\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442263 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b287991-894d-4985-b0e0-74a46ae4a7d3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442286 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b287991-894d-4985-b0e0-74a46ae4a7d3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442302 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442334 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add58d5d-445b-47e0-9399-180a0c4571a8-config\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442380 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gj8k\" (UniqueName: \"kubernetes.io/projected/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-kube-api-access-9gj8k\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.442416 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/add58d5d-445b-47e0-9399-180a0c4571a8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.441977 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-n26vk"] Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.445825 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:53.945790232 +0000 UTC m=+141.935019029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.457020 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.501166 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" event={"ID":"416da0ef-bf05-4226-aae9-692b0ccf210d","Type":"ContainerStarted","Data":"ed191bcfa3925549a4fd9f50b3c4d8e11ae838d74bbebd263059b46a07faa49c"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.501875 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.522185 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" Oct 01 09:11:53 crc kubenswrapper[4983]: W1001 09:11:53.538886 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7c24919_2815_458f_af2e_11ca53444cc1.slice/crio-cf957dd37e99c8046bf271a0b495abdf8ae708fc1339027d0604910e9a2873c9 WatchSource:0}: Error finding container cf957dd37e99c8046bf271a0b495abdf8ae708fc1339027d0604910e9a2873c9: Status 404 returned error can't find the container with id cf957dd37e99c8046bf271a0b495abdf8ae708fc1339027d0604910e9a2873c9 Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545210 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.545323 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.045300431 +0000 UTC m=+142.034529228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545465 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6deda0d-2483-4468-9959-7d7b47dd731d-proxy-tls\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545486 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqbk\" (UniqueName: \"kubernetes.io/projected/7330817e-3c60-4eaa-b546-76a64214a03f-kube-api-access-sgqbk\") pod \"migrator-59844c95c7-n5stn\" (UID: \"7330817e-3c60-4eaa-b546-76a64214a03f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545602 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjn7n\" (UniqueName: \"kubernetes.io/projected/a6deda0d-2483-4468-9959-7d7b47dd731d-kube-api-access-xjn7n\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545639 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-default-certificate\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545657 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-stats-auth\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545677 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-csi-data-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545728 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545753 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2df46f36-de11-4b06-86d4-7ad0e6e58737-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545776 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b287991-894d-4985-b0e0-74a46ae4a7d3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545881 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc3891f-6cc8-445d-b852-6b34df2bf821-config\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545958 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-config\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.545995 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72kv\" (UniqueName: \"kubernetes.io/projected/11a442c1-aca4-467a-bbb7-5ccea25def0b-kube-api-access-r72kv\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546013 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6ln\" (UniqueName: \"kubernetes.io/projected/5b507fa0-037d-4a4d-bde6-9485e8735942-kube-api-access-5r6ln\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546076 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b507fa0-037d-4a4d-bde6-9485e8735942-cert\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546156 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-registration-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546173 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-metrics-certs\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546188 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546229 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b287991-894d-4985-b0e0-74a46ae4a7d3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546245 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11a442c1-aca4-467a-bbb7-5ccea25def0b-service-ca-bundle\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546261 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-mountpoint-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546303 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b287991-894d-4985-b0e0-74a46ae4a7d3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546318 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-socket-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546388 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546440 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add58d5d-445b-47e0-9399-180a0c4571a8-config\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546457 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbv8p\" (UniqueName: \"kubernetes.io/projected/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-kube-api-access-zbv8p\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546548 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-plugins-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546584 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gj8k\" (UniqueName: \"kubernetes.io/projected/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-kube-api-access-9gj8k\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546629 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/add58d5d-445b-47e0-9399-180a0c4571a8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546682 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6bc3891f-6cc8-445d-b852-6b34df2bf821-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546699 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-certs\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546744 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-node-bootstrap-token\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546798 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwmbn\" (UniqueName: \"kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546863 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6deda0d-2483-4468-9959-7d7b47dd731d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546886 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/add58d5d-445b-47e0-9399-180a0c4571a8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546944 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-serving-cert\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.546962 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qch66\" (UniqueName: \"kubernetes.io/projected/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-kube-api-access-qch66\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.547050 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2df46f36-de11-4b06-86d4-7ad0e6e58737-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.547077 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcpmh\" (UniqueName: \"kubernetes.io/projected/2df46f36-de11-4b06-86d4-7ad0e6e58737-kube-api-access-kcpmh\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.547093 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bc3891f-6cc8-445d-b852-6b34df2bf821-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.554844 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc3891f-6cc8-445d-b852-6b34df2bf821-config\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.558554 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add58d5d-445b-47e0-9399-180a0c4571a8-config\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.560568 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.0605529 +0000 UTC m=+142.049781787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.561858 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.563592 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-config\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.586964 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2df46f36-de11-4b06-86d4-7ad0e6e58737-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.587554 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b287991-894d-4985-b0e0-74a46ae4a7d3-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.588204 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-stats-auth\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.589346 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6deda0d-2483-4468-9959-7d7b47dd731d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.590009 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6deda0d-2483-4468-9959-7d7b47dd731d-proxy-tls\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.597923 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.606430 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-default-certificate\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.625921 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b287991-894d-4985-b0e0-74a46ae4a7d3-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.627232 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.639333 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11a442c1-aca4-467a-bbb7-5ccea25def0b-service-ca-bundle\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.640180 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2df46f36-de11-4b06-86d4-7ad0e6e58737-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.645403 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" event={"ID":"3b1a6a6e-c4f8-4e5e-acbf-62220edee276","Type":"ContainerStarted","Data":"941d91357d03537eb737e79a923af191749a0da0ac770fa3d179b4e530698b59"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.646535 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/11a442c1-aca4-467a-bbb7-5ccea25def0b-metrics-certs\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.648964 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.650318 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-certs\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.651012 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-node-bootstrap-token\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.651900 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.151659694 +0000 UTC m=+142.140888531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652080 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-csi-data-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652149 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652206 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6ln\" (UniqueName: \"kubernetes.io/projected/5b507fa0-037d-4a4d-bde6-9485e8735942-kube-api-access-5r6ln\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652237 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b507fa0-037d-4a4d-bde6-9485e8735942-cert\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652271 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-registration-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652300 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-mountpoint-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652338 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-socket-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652368 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbv8p\" (UniqueName: \"kubernetes.io/projected/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-kube-api-access-zbv8p\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652725 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-csi-data-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.652926 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6bc3891f-6cc8-445d-b852-6b34df2bf821-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.653035 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.153026718 +0000 UTC m=+142.142255505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.653486 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-mountpoint-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.653591 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-registration-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.655594 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-socket-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.657604 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.658147 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bc3891f-6cc8-445d-b852-6b34df2bf821-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvtld\" (UID: \"6bc3891f-6cc8-445d-b852-6b34df2bf821\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.658294 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/add58d5d-445b-47e0-9399-180a0c4571a8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.658922 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72kv\" (UniqueName: \"kubernetes.io/projected/11a442c1-aca4-467a-bbb7-5ccea25def0b-kube-api-access-r72kv\") pod \"router-default-5444994796-5wzgd\" (UID: \"11a442c1-aca4-467a-bbb7-5ccea25def0b\") " pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.662295 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" event={"ID":"c8199b4d-1870-489d-a48e-d8a0376f9090","Type":"ContainerStarted","Data":"97765f8957684705df9d9c849642dbb7e1a15af76e8278b7c79ac2cb668a7136"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.662340 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" event={"ID":"c8199b4d-1870-489d-a48e-d8a0376f9090","Type":"ContainerStarted","Data":"c185f27b621eba55bbb4d083207bf2958cb4c0fb64d1b2a0890a48e6e0066ea4"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.684230 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b287991-894d-4985-b0e0-74a46ae4a7d3-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-76knj\" (UID: \"7b287991-894d-4985-b0e0-74a46ae4a7d3\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.688823 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-serving-cert\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.694509 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jjb8k" event={"ID":"46bfb1e9-ebaf-4e47-9983-71e6d6a01154","Type":"ContainerStarted","Data":"bd126f4209b9b294318f69ba1e81141958613fc6c17fdb7459abbafc185a880a"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.694550 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jjb8k" event={"ID":"46bfb1e9-ebaf-4e47-9983-71e6d6a01154","Type":"ContainerStarted","Data":"8c557674a1dadf2efc647eabb562cfd49386f9cbef3eb306e5288a0657413f90"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.696371 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-node-bootstrap-token\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.698474 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.699281 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b507fa0-037d-4a4d-bde6-9485e8735942-cert\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.700016 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" event={"ID":"72a01b33-7f81-470b-8c5c-1a2d12f40420","Type":"ContainerStarted","Data":"aaddb414ff54155d038b690b8f01d9715176cb7b0cb3688ac8d0713ecab1ed32"} Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.700623 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f"] Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.702961 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqbk\" (UniqueName: \"kubernetes.io/projected/7330817e-3c60-4eaa-b546-76a64214a03f-kube-api-access-sgqbk\") pod \"migrator-59844c95c7-n5stn\" (UID: \"7330817e-3c60-4eaa-b546-76a64214a03f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.703326 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.704491 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-certs\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.716542 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-jjb8k container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.716620 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jjb8k" podUID="46bfb1e9-ebaf-4e47-9983-71e6d6a01154" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.725782 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjn7n\" (UniqueName: \"kubernetes.io/projected/a6deda0d-2483-4468-9959-7d7b47dd731d-kube-api-access-xjn7n\") pod \"machine-config-controller-84d6567774-5rckn\" (UID: \"a6deda0d-2483-4468-9959-7d7b47dd731d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.726502 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwmbn\" (UniqueName: \"kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn\") pod \"collect-profiles-29321820-vlk57\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.727723 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl"] Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.731365 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/add58d5d-445b-47e0-9399-180a0c4571a8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fvqlm\" (UID: \"add58d5d-445b-47e0-9399-180a0c4571a8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.731504 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.746544 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.753576 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.753746 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-plugins-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.753959 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qch66\" (UniqueName: \"kubernetes.io/projected/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-kube-api-access-qch66\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.755026 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.25501049 +0000 UTC m=+142.244239287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.756106 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-plugins-dir\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.757361 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.772383 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.780017 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gj8k\" (UniqueName: \"kubernetes.io/projected/cf476870-f3ea-4e49-bb77-4d3f7edab9ba-kube-api-access-9gj8k\") pod \"service-ca-operator-777779d784-rj8kt\" (UID: \"cf476870-f3ea-4e49-bb77-4d3f7edab9ba\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.797335 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcpmh\" (UniqueName: \"kubernetes.io/projected/2df46f36-de11-4b06-86d4-7ad0e6e58737-kube-api-access-kcpmh\") pod \"openshift-controller-manager-operator-756b6f6bc6-tgkqm\" (UID: \"2df46f36-de11-4b06-86d4-7ad0e6e58737\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.846330 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbv8p\" (UniqueName: \"kubernetes.io/projected/cc9f302b-e3db-48ed-a10b-00e17c3d5b6d-kube-api-access-zbv8p\") pod \"machine-config-server-97dph\" (UID: \"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d\") " pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.847631 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.848604 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.855333 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.855612 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.355600356 +0000 UTC m=+142.344829153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.856701 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6ln\" (UniqueName: \"kubernetes.io/projected/5b507fa0-037d-4a4d-bde6-9485e8735942-kube-api-access-5r6ln\") pod \"ingress-canary-jzscr\" (UID: \"5b507fa0-037d-4a4d-bde6-9485e8735942\") " pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.864507 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-97dph" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.873464 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.891411 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jzscr" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.892360 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qch66\" (UniqueName: \"kubernetes.io/projected/8501eee0-e9f1-4b16-ba33-f8ba28add8fd-kube-api-access-qch66\") pod \"csi-hostpathplugin-92fqg\" (UID: \"8501eee0-e9f1-4b16-ba33-f8ba28add8fd\") " pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.916128 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.939053 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.947800 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.955975 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:53 crc kubenswrapper[4983]: E1001 09:11:53.956503 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.456482759 +0000 UTC m=+142.445711556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:53 crc kubenswrapper[4983]: I1001 09:11:53.980162 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.058964 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.059435 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.559420635 +0000 UTC m=+142.548649432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.169085 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.169986 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.669970375 +0000 UTC m=+142.659199172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.192978 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.277036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.277595 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.77758269 +0000 UTC m=+142.766811487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.346337 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" podStartSLOduration=122.346322923 podStartE2EDuration="2m2.346322923s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:54.345082291 +0000 UTC m=+142.334311088" watchObservedRunningTime="2025-10-01 09:11:54.346322923 +0000 UTC m=+142.335551720" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.379952 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.380130 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.880102025 +0000 UTC m=+142.869330822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.380390 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.380718 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.880706071 +0000 UTC m=+142.869934868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.391657 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-tf547" podStartSLOduration=122.391636189 podStartE2EDuration="2m2.391636189s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:54.38893332 +0000 UTC m=+142.378162137" watchObservedRunningTime="2025-10-01 09:11:54.391636189 +0000 UTC m=+142.380864986" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.482041 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.482485 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:54.982470756 +0000 UTC m=+142.971699543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.587117 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.587799 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.087788552 +0000 UTC m=+143.077017349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.626640 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m"] Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.649152 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w5pt5"] Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.652526 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgxgp"] Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.683733 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.691106 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.691339 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.191325383 +0000 UTC m=+143.180554180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.706893 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5wzgd" event={"ID":"11a442c1-aca4-467a-bbb7-5ccea25def0b","Type":"ContainerStarted","Data":"bef308deeed6388fb3ed78eb54b2bdd37369b444d9dd1bde8b846b17690a22c3"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.707134 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5wzgd" event={"ID":"11a442c1-aca4-467a-bbb7-5ccea25def0b","Type":"ContainerStarted","Data":"23a3715402ae8d3af8ae86d93b2311f0fd7e915329c7668076a8b4ac615fa9b7"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.719968 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-n26vk" event={"ID":"f7c24919-2815-458f-af2e-11ca53444cc1","Type":"ContainerStarted","Data":"4ba0a0a9eb8ae9f5b71188ae24758b27da54f344f0bc850e81aa9ed583846d84"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.720011 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.720022 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-n26vk" event={"ID":"f7c24919-2815-458f-af2e-11ca53444cc1","Type":"ContainerStarted","Data":"cf957dd37e99c8046bf271a0b495abdf8ae708fc1339027d0604910e9a2873c9"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.720032 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" event={"ID":"e5f07bb1-dbb5-4a9c-90ed-caec428cd247","Type":"ContainerStarted","Data":"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.720041 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" event={"ID":"e5f07bb1-dbb5-4a9c-90ed-caec428cd247","Type":"ContainerStarted","Data":"44c91d8db86a9c701f7b0ae9a8d3baa00639d6f7d65a8047bb9e35632dc10e74"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.732917 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" event={"ID":"027246b3-e611-429e-80c0-653dbde84d0a","Type":"ContainerStarted","Data":"e589bec046cdf6d0a6b424763032ac0c1340de7057b5a60a48983e939e6abe39"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.732961 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" event={"ID":"027246b3-e611-429e-80c0-653dbde84d0a","Type":"ContainerStarted","Data":"e6999904204150cc7a2da8ba4efbe05ff51b9abb81f697a96e9bc04de2e0c97d"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.733233 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.737252 4983 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-z72rd container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.25:6443/healthz\": dial tcp 10.217.0.25:6443: connect: connection refused" start-of-body= Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.737298 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.25:6443/healthz\": dial tcp 10.217.0.25:6443: connect: connection refused" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.738688 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" event={"ID":"72a01b33-7f81-470b-8c5c-1a2d12f40420","Type":"ContainerStarted","Data":"2c5f3aa7056ba34fe75ffaecd94ee12ba94da5c192d796145c41e92c4d43ffdf"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.739722 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-97dph" event={"ID":"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d","Type":"ContainerStarted","Data":"ec578f990ac8cae3cff2997d64fe2dd11db37cdbf5c05a0a15ba8d72c686609b"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.739744 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-97dph" event={"ID":"cc9f302b-e3db-48ed-a10b-00e17c3d5b6d","Type":"ContainerStarted","Data":"0e48781e42cc3250a011174f1f085ea6e788bdbf1ce59d45e178e4ff3678ab8b"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.752150 4983 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tg26f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.752202 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" podUID="027246b3-e611-429e-80c0-653dbde84d0a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.764879 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.792049 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.793634 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.293614872 +0000 UTC m=+143.282843689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.802669 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" podStartSLOduration=122.802652183 podStartE2EDuration="2m2.802652183s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:54.773869569 +0000 UTC m=+142.763098366" watchObservedRunningTime="2025-10-01 09:11:54.802652183 +0000 UTC m=+142.791880980" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.807519 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" event={"ID":"afb7d248-4e64-47c3-a974-10a00e1f2fca","Type":"ContainerStarted","Data":"724e750396d19f9f903b35232224d7ab86223440c8f765ddaf4df043ddb0e5f2"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.807562 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" event={"ID":"afb7d248-4e64-47c3-a974-10a00e1f2fca","Type":"ContainerStarted","Data":"8891fcded637489af11e762dd226c2d1d1260386b6e3bb59eda3286940bb327a"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.829123 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.829175 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.848448 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" event={"ID":"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d","Type":"ContainerStarted","Data":"99590a893ddd196c7fcd4f4d4dbce4bbdc1fe44f4d88ccdcb92246ffa7356324"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.848488 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" event={"ID":"7f6005e9-43e4-4cf2-9428-fb48a2a26f4d","Type":"ContainerStarted","Data":"2db9bbdf740e523c79eb2dfa680019550c91ad4321998453a3d33c6901474879"} Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.849017 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-jjb8k container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.849083 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jjb8k" podUID="46bfb1e9-ebaf-4e47-9983-71e6d6a01154" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.874821 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-jjb8k" podStartSLOduration=122.874784583 podStartE2EDuration="2m2.874784583s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:54.874730552 +0000 UTC m=+142.863959359" watchObservedRunningTime="2025-10-01 09:11:54.874784583 +0000 UTC m=+142.864013380" Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.896169 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:54 crc kubenswrapper[4983]: E1001 09:11:54.951595 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.451574092 +0000 UTC m=+143.440802889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:54 crc kubenswrapper[4983]: I1001 09:11:54.970969 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jnpd"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.001533 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" podStartSLOduration=123.001515666 podStartE2EDuration="2m3.001515666s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:54.986288848 +0000 UTC m=+142.975517655" watchObservedRunningTime="2025-10-01 09:11:55.001515666 +0000 UTC m=+142.990744463" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.015684 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.016110 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.516097208 +0000 UTC m=+143.505325995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.060360 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" podStartSLOduration=123.060342907 podStartE2EDuration="2m3.060342907s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.043141288 +0000 UTC m=+143.032370085" watchObservedRunningTime="2025-10-01 09:11:55.060342907 +0000 UTC m=+143.049571704" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.122149 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.123035 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.623020366 +0000 UTC m=+143.612249163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.145068 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6brq6" podStartSLOduration=123.145050257 podStartE2EDuration="2m3.145050257s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.144286228 +0000 UTC m=+143.133515025" watchObservedRunningTime="2025-10-01 09:11:55.145050257 +0000 UTC m=+143.134279054" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.223913 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.224185 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.724173766 +0000 UTC m=+143.713402553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.324829 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.325461 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.825445699 +0000 UTC m=+143.814674496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.318563 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-284pv" podStartSLOduration=123.318545513 podStartE2EDuration="2m3.318545513s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.312507789 +0000 UTC m=+143.301736586" watchObservedRunningTime="2025-10-01 09:11:55.318545513 +0000 UTC m=+143.307774310" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.355687 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.369554 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.425090 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fvslk"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.426593 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.427049 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:55.92702717 +0000 UTC m=+143.916256037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.457912 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.468112 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rpg24"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.522164 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" podStartSLOduration=123.522148207 podStartE2EDuration="2m3.522148207s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.500403632 +0000 UTC m=+143.489632439" watchObservedRunningTime="2025-10-01 09:11:55.522148207 +0000 UTC m=+143.511377004" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.523682 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.527212 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.527545 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.027530804 +0000 UTC m=+144.016759601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.570494 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.591143 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-49krl" podStartSLOduration=123.591125926 podStartE2EDuration="2m3.591125926s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.586244202 +0000 UTC m=+143.575472999" watchObservedRunningTime="2025-10-01 09:11:55.591125926 +0000 UTC m=+143.580354723" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.600069 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wrsrk"] Oct 01 09:11:55 crc kubenswrapper[4983]: W1001 09:11:55.613563 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7330817e_3c60_4eaa_b546_76a64214a03f.slice/crio-5311cfb8f8ceb10a3ef79f6a3b375e6749eb36bf797889492f803b5886c2f975 WatchSource:0}: Error finding container 5311cfb8f8ceb10a3ef79f6a3b375e6749eb36bf797889492f803b5886c2f975: Status 404 returned error can't find the container with id 5311cfb8f8ceb10a3ef79f6a3b375e6749eb36bf797889492f803b5886c2f975 Oct 01 09:11:55 crc kubenswrapper[4983]: W1001 09:11:55.621781 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod177c9b4a_8758_47be_9b2e_004485032310.slice/crio-bfeeaa3dfad8d709f49aff964ec6b90f194ffca60a197ea90484f23292045556 WatchSource:0}: Error finding container bfeeaa3dfad8d709f49aff964ec6b90f194ffca60a197ea90484f23292045556: Status 404 returned error can't find the container with id bfeeaa3dfad8d709f49aff964ec6b90f194ffca60a197ea90484f23292045556 Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.636314 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.636680 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.136665828 +0000 UTC m=+144.125894625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.721585 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.733297 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.737895 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.738850 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.238833604 +0000 UTC m=+144.228062401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.739514 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-92fqg"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.741538 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5wzgd" podStartSLOduration=123.741520252 podStartE2EDuration="2m3.741520252s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.73749374 +0000 UTC m=+143.726722537" watchObservedRunningTime="2025-10-01 09:11:55.741520252 +0000 UTC m=+143.730749049" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.758059 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.768982 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:11:55 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:11:55 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:11:55 crc kubenswrapper[4983]: healthz check failed Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.769041 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.820226 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-97dph" podStartSLOduration=5.82021018 podStartE2EDuration="5.82021018s" podCreationTimestamp="2025-10-01 09:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.819587403 +0000 UTC m=+143.808816220" watchObservedRunningTime="2025-10-01 09:11:55.82021018 +0000 UTC m=+143.809438977" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.840031 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.840428 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.340414635 +0000 UTC m=+144.329643442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.847901 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.884480 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.910537 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-gwghb" podStartSLOduration=123.910520623 podStartE2EDuration="2m3.910520623s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:55.908459901 +0000 UTC m=+143.897688698" watchObservedRunningTime="2025-10-01 09:11:55.910520623 +0000 UTC m=+143.899749420" Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.921900 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" event={"ID":"65fcd0cc-38b3-4951-8379-261a96372b6f","Type":"ContainerStarted","Data":"3cdb2751eac505427aad2ede249253cdbdccb492ad7d5aa73a91838eba22bd3f"} Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.921938 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" event={"ID":"65fcd0cc-38b3-4951-8379-261a96372b6f","Type":"ContainerStarted","Data":"aeaf15ada20f0306fb4afb6d0f5aa8a30f42c3065f0e693d60adb85973fd8fc7"} Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.921964 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" event={"ID":"65fcd0cc-38b3-4951-8379-261a96372b6f","Type":"ContainerStarted","Data":"d150a275b351e423d20bfb05eb37a74685b66b975e399e34301f78766e55419c"} Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.922636 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.936374 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.936610 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jzscr"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.943194 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:55 crc kubenswrapper[4983]: E1001 09:11:55.944196 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.444159701 +0000 UTC m=+144.433388558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.945380 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.946657 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm"] Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.977017 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" event={"ID":"72a01b33-7f81-470b-8c5c-1a2d12f40420","Type":"ContainerStarted","Data":"409e3bdac00d1d4db49e6aa113f24d83104e22b7a4a3600dd2abe4b36ff3cab1"} Oct 01 09:11:55 crc kubenswrapper[4983]: I1001 09:11:55.994293 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" event={"ID":"59073ee5-3101-46ac-88b8-ad59cdb2f5a8","Type":"ContainerStarted","Data":"644644bca46dd28cbb51ff853c61e4ab88d6f62dad8ed692bb0a7c4c8bceef47"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.006049 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" event={"ID":"dd4036a5-2db8-4cec-98c1-7723eebcfdcd","Type":"ContainerStarted","Data":"2ad094c821c7b4ee49646af0c638e1ef02bc544e5093a90baaf2f954bf056e89"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.006094 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" event={"ID":"dd4036a5-2db8-4cec-98c1-7723eebcfdcd","Type":"ContainerStarted","Data":"780611b66f22377203c370d75cc4f56e8867b06bcf59070e87a9f63c12a1cca9"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.006107 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" event={"ID":"dd4036a5-2db8-4cec-98c1-7723eebcfdcd","Type":"ContainerStarted","Data":"3909e3baf7d8ae25ef7e3ae9631aa08c27acb1713a4d796bd2152378d15ce430"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.006339 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.007692 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" event={"ID":"40ccad2f-114b-43c1-8b78-d5c252dea8ff","Type":"ContainerStarted","Data":"960fb05200c6b763c18e99a08c253bf554452da41fbd01ed19b4c60b8cbd2f03"} Oct 01 09:11:56 crc kubenswrapper[4983]: W1001 09:11:56.024767 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b287991_894d_4985_b0e0_74a46ae4a7d3.slice/crio-0d3b3f4895d6c365b065a377e3e69318e33316d5562aa7598710e3a1e75bb901 WatchSource:0}: Error finding container 0d3b3f4895d6c365b065a377e3e69318e33316d5562aa7598710e3a1e75bb901: Status 404 returned error can't find the container with id 0d3b3f4895d6c365b065a377e3e69318e33316d5562aa7598710e3a1e75bb901 Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.029108 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" podStartSLOduration=124.029088208 podStartE2EDuration="2m4.029088208s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.024617484 +0000 UTC m=+144.013846281" watchObservedRunningTime="2025-10-01 09:11:56.029088208 +0000 UTC m=+144.018317005" Oct 01 09:11:56 crc kubenswrapper[4983]: W1001 09:11:56.039949 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf476870_f3ea_4e49_bb77_4d3f7edab9ba.slice/crio-7df694689e91103a62f15c4ab5f77ce3ca66bc54fe7e0853a2bf3d9a9fd6cda7 WatchSource:0}: Error finding container 7df694689e91103a62f15c4ab5f77ce3ca66bc54fe7e0853a2bf3d9a9fd6cda7: Status 404 returned error can't find the container with id 7df694689e91103a62f15c4ab5f77ce3ca66bc54fe7e0853a2bf3d9a9fd6cda7 Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.048120 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.049875 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.549861638 +0000 UTC m=+144.539090435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.051478 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" event={"ID":"7330817e-3c60-4eaa-b546-76a64214a03f","Type":"ContainerStarted","Data":"5311cfb8f8ceb10a3ef79f6a3b375e6749eb36bf797889492f803b5886c2f975"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.057697 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" event={"ID":"db154c97-53b6-4602-a6ab-4425c548f724","Type":"ContainerStarted","Data":"7b2a825c102d59ce4a1875a5b05f9008d6a29fa875b24071421e2d222aca3247"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.060217 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" podStartSLOduration=124.060200371 podStartE2EDuration="2m4.060200371s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.059634557 +0000 UTC m=+144.048863354" watchObservedRunningTime="2025-10-01 09:11:56.060200371 +0000 UTC m=+144.049429168" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.064613 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" event={"ID":"01cf3320-f019-4bfb-9de1-826d0335ec36","Type":"ContainerStarted","Data":"95070bbcbdcb93840fb8c71c546e0fea6c5efe343f8ceb6930d6f8c9c5fa8a0b"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.064656 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" event={"ID":"01cf3320-f019-4bfb-9de1-826d0335ec36","Type":"ContainerStarted","Data":"3762b809a1d716cf9d812a21ee6aa564374724686cf672943d146f75a46aa07a"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.064828 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.066117 4983 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4xzjs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.066148 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" podUID="01cf3320-f019-4bfb-9de1-826d0335ec36" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.070455 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" event={"ID":"add58d5d-445b-47e0-9399-180a0c4571a8","Type":"ContainerStarted","Data":"e136673337d8393fd58c628f2034fc6a7991114c80dfec8207fdd2e7bf05eaef"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.071631 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" event={"ID":"a199e85c-98a8-4563-81a6-196998a19005","Type":"ContainerStarted","Data":"3585414e687f49642614de54551b6dbfc57f3c318f4752fd0415498543390460"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.071671 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" event={"ID":"a199e85c-98a8-4563-81a6-196998a19005","Type":"ContainerStarted","Data":"3ed738223af0fc9a61d526e6bd2f8f0feaf2456be8a3e3e9d07d9f157cb56132"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.072488 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.074024 4983 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xhsfh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.074062 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.075505 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" event={"ID":"05d0252d-cab0-454c-aa99-fe8f77e658bc","Type":"ContainerStarted","Data":"fd31bf8f06edd5edaeed48e7435a16e75f6e70235b7f47bc0a4f7f784274f267"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.077887 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" event={"ID":"e13daed8-fbbc-4c2e-93fc-411c659910e9","Type":"ContainerStarted","Data":"c24de8f72eb12de4b031d397a5319e8904e9979dbbd8c7fa656a58fd88a29fcd"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.079352 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" event={"ID":"9700b45d-c305-4dae-848f-c05699323db2","Type":"ContainerStarted","Data":"a8f417c26dfa4d7b439261149e3c1edc1cff7e61e478b88be05af4ae17b7f4fe"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.079399 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" event={"ID":"9700b45d-c305-4dae-848f-c05699323db2","Type":"ContainerStarted","Data":"f5666be905c08d511a5783ada8c0e55e8f1304f626c001e4cc9b8d89637a074e"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.079409 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" event={"ID":"9700b45d-c305-4dae-848f-c05699323db2","Type":"ContainerStarted","Data":"6638959f5ba894fbc85eda7ad6550913e3dc51edf44921a56fcba892b7e2f471"} Oct 01 09:11:56 crc kubenswrapper[4983]: W1001 09:11:56.080638 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b507fa0_037d_4a4d_bde6_9485e8735942.slice/crio-947bdf4718e32fc5cec04a0d423af16966ab0237d31411ee6f1316ee59106492 WatchSource:0}: Error finding container 947bdf4718e32fc5cec04a0d423af16966ab0237d31411ee6f1316ee59106492: Status 404 returned error can't find the container with id 947bdf4718e32fc5cec04a0d423af16966ab0237d31411ee6f1316ee59106492 Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.086954 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" event={"ID":"61c94092-2bfe-4a9a-ab80-c6268da4d42c","Type":"ContainerStarted","Data":"25370f1b03265e1f96df5e935c7ba046a77a4c67cf88d433d4a1f68f5b4e9bb9"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.087010 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" event={"ID":"61c94092-2bfe-4a9a-ab80-c6268da4d42c","Type":"ContainerStarted","Data":"682fce06c65688422993f22e83dd53f69780060bad521b13f4549b9653b54f0d"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.090573 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" event={"ID":"1cdabf47-3a38-4c24-b2e8-40ae2f062f08","Type":"ContainerStarted","Data":"599de3b110f89a6e2be047a5ad736ec0214179e6e9ac896c104f95042bc67541"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.090635 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" event={"ID":"1cdabf47-3a38-4c24-b2e8-40ae2f062f08","Type":"ContainerStarted","Data":"ea33e43f1a9205048e0d0f59c36319146ae1765a815347a45727940e5030be7d"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.098523 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.113124 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wrsrk" event={"ID":"177c9b4a-8758-47be-9b2e-004485032310","Type":"ContainerStarted","Data":"bfeeaa3dfad8d709f49aff964ec6b90f194ffca60a197ea90484f23292045556"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.139235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" event={"ID":"8501eee0-e9f1-4b16-ba33-f8ba28add8fd","Type":"ContainerStarted","Data":"10d48839c2ec5874fbd1f961f06227448436effd9a6752ff4f5b35628e4abd18"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.141353 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" event={"ID":"50427e6c-1ddf-4141-9d2d-24ac5726c401","Type":"ContainerStarted","Data":"8fc22bbfde5a859fb8d3ca30ac16ffad0037a9895f51e875a36c54bbe2dcc48c"} Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.143676 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.147605 4983 patch_prober.go:28] interesting pod/console-operator-58897d9998-n26vk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.147651 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-n26vk" podUID="f7c24919-2815-458f-af2e-11ca53444cc1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.149245 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.150388 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.650371101 +0000 UTC m=+144.639599948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.156000 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4bh2" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.183318 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" podStartSLOduration=124.183296041 podStartE2EDuration="2m4.183296041s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.177483783 +0000 UTC m=+144.166712570" watchObservedRunningTime="2025-10-01 09:11:56.183296041 +0000 UTC m=+144.172524838" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.217716 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-w5pt5" podStartSLOduration=124.217698029 podStartE2EDuration="2m4.217698029s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.213291756 +0000 UTC m=+144.202520543" watchObservedRunningTime="2025-10-01 09:11:56.217698029 +0000 UTC m=+144.206926826" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.250638 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.252408 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.752393673 +0000 UTC m=+144.741622470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.297710 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" podStartSLOduration=124.29769295 podStartE2EDuration="2m4.29769295s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.297672499 +0000 UTC m=+144.286901296" watchObservedRunningTime="2025-10-01 09:11:56.29769295 +0000 UTC m=+144.286921747" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.302101 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" podStartSLOduration=124.302085702 podStartE2EDuration="2m4.302085702s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.258099069 +0000 UTC m=+144.247327866" watchObservedRunningTime="2025-10-01 09:11:56.302085702 +0000 UTC m=+144.291314499" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.334068 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rgz6d" podStartSLOduration=125.334052957 podStartE2EDuration="2m5.334052957s" podCreationTimestamp="2025-10-01 09:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.33144041 +0000 UTC m=+144.320669207" watchObservedRunningTime="2025-10-01 09:11:56.334052957 +0000 UTC m=+144.323281754" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.352585 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.353730 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.853709539 +0000 UTC m=+144.842938336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.455796 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.456144 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:56.956133071 +0000 UTC m=+144.945361868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.466532 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" podStartSLOduration=124.466506695 podStartE2EDuration="2m4.466506695s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.465650764 +0000 UTC m=+144.454879571" watchObservedRunningTime="2025-10-01 09:11:56.466506695 +0000 UTC m=+144.455735492" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.467123 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" podStartSLOduration=124.467119591 podStartE2EDuration="2m4.467119591s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.428586378 +0000 UTC m=+144.417815195" watchObservedRunningTime="2025-10-01 09:11:56.467119591 +0000 UTC m=+144.456348378" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.506937 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.559320 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.559597 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.05958024 +0000 UTC m=+145.048809037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.559625 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.559921 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.059914288 +0000 UTC m=+145.049143085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.559907 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgxgp" podStartSLOduration=124.559846316 podStartE2EDuration="2m4.559846316s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.499575519 +0000 UTC m=+144.488804326" watchObservedRunningTime="2025-10-01 09:11:56.559846316 +0000 UTC m=+144.549075113" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.561172 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" podStartSLOduration=124.56115891 podStartE2EDuration="2m4.56115891s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.554558491 +0000 UTC m=+144.543787308" watchObservedRunningTime="2025-10-01 09:11:56.56115891 +0000 UTC m=+144.550387707" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.660852 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.661097 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.161067568 +0000 UTC m=+145.150296365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.661310 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.661580 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.161568662 +0000 UTC m=+145.150797459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.691580 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-n26vk" podStartSLOduration=124.691562016 podStartE2EDuration="2m4.691562016s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:56.687998475 +0000 UTC m=+144.677227282" watchObservedRunningTime="2025-10-01 09:11:56.691562016 +0000 UTC m=+144.680790813" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.696436 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tg26f" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.763392 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.763693 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.263677296 +0000 UTC m=+145.252906093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.766648 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:11:56 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:11:56 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:11:56 crc kubenswrapper[4983]: healthz check failed Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.766698 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.775401 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.776753 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.802156 4983 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bkjc6 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]log ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]etcd ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/generic-apiserver-start-informers ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/max-in-flight-filter ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 01 09:11:56 crc kubenswrapper[4983]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 01 09:11:56 crc kubenswrapper[4983]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/project.openshift.io-projectcache ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-startinformers ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 01 09:11:56 crc kubenswrapper[4983]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 01 09:11:56 crc kubenswrapper[4983]: livez check failed Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.802525 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" podUID="c8199b4d-1870-489d-a48e-d8a0376f9090" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.867726 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.873875 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.373849496 +0000 UTC m=+145.363078293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.880333 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.881306 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.894712 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:56 crc kubenswrapper[4983]: I1001 09:11:56.970640 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:56 crc kubenswrapper[4983]: E1001 09:11:56.972028 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.47201265 +0000 UTC m=+145.461241447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.072912 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.073302 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.573285054 +0000 UTC m=+145.562513851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.147496 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" event={"ID":"cf476870-f3ea-4e49-bb77-4d3f7edab9ba","Type":"ContainerStarted","Data":"0262f6fb0423dd5e0066ae7493b917a16e221388da70f2e5aad6e782d2099f4e"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.147553 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" event={"ID":"cf476870-f3ea-4e49-bb77-4d3f7edab9ba","Type":"ContainerStarted","Data":"7df694689e91103a62f15c4ab5f77ce3ca66bc54fe7e0853a2bf3d9a9fd6cda7"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.150970 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" event={"ID":"50427e6c-1ddf-4141-9d2d-24ac5726c401","Type":"ContainerStarted","Data":"de499b577becce000b312d79dfcbcdf3ba87789a2599ffd3528d7e8e6e674e01"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.152486 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mjxft" event={"ID":"59073ee5-3101-46ac-88b8-ad59cdb2f5a8","Type":"ContainerStarted","Data":"f413d285f31a9f62208955fb25c66e7696f81194de4205e21e13eac11e9b17b6"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.153835 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" event={"ID":"add58d5d-445b-47e0-9399-180a0c4571a8","Type":"ContainerStarted","Data":"2afa96e10b203784a091d5f4416559d2b662c9a45928abb3388b4103aec9258c"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.154985 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5svzw" event={"ID":"05d0252d-cab0-454c-aa99-fe8f77e658bc","Type":"ContainerStarted","Data":"eaa472d229ae4222b34d04e468647a50fe027b50fb9d3d90a7a48988a5e0205e"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.156078 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jzscr" event={"ID":"5b507fa0-037d-4a4d-bde6-9485e8735942","Type":"ContainerStarted","Data":"1afabe4c8eccc9338ab2cbf4bc2915a88442bc9a291020a7e698f503f6f9d206"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.156107 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jzscr" event={"ID":"5b507fa0-037d-4a4d-bde6-9485e8735942","Type":"ContainerStarted","Data":"947bdf4718e32fc5cec04a0d423af16966ab0237d31411ee6f1316ee59106492"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.157597 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fvslk" event={"ID":"db154c97-53b6-4602-a6ab-4425c548f724","Type":"ContainerStarted","Data":"7a51398c81183445672dfe96bc7593772d3046b8f34cbc3b51930aaff370eadc"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.161582 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rj8kt" podStartSLOduration=125.161566016 podStartE2EDuration="2m5.161566016s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.158868357 +0000 UTC m=+145.148097154" watchObservedRunningTime="2025-10-01 09:11:57.161566016 +0000 UTC m=+145.150794813" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.161865 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" event={"ID":"a6deda0d-2483-4468-9959-7d7b47dd731d","Type":"ContainerStarted","Data":"1245b43c0488153ad166f1bf7aafa1ac2bdc927898b802a3adaab56a97092fcb"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.161899 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" event={"ID":"a6deda0d-2483-4468-9959-7d7b47dd731d","Type":"ContainerStarted","Data":"4a6acff028b4298f0cc29b0ac202f7a268d8c5c45efe78d0b239d59c03099532"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.161911 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" event={"ID":"a6deda0d-2483-4468-9959-7d7b47dd731d","Type":"ContainerStarted","Data":"5d3d781a10e065ae5d57f4f2930656f7edbe00ae754ca43dc3dce2bdbbc30891"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.164968 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" event={"ID":"e13daed8-fbbc-4c2e-93fc-411c659910e9","Type":"ContainerStarted","Data":"19134cd1b0cfe5908e94521f12abb54503a61ccfcfeeb10a2b4e532fb31fe36e"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.165701 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.166488 4983 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5mdqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.166522 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" podUID="e13daed8-fbbc-4c2e-93fc-411c659910e9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.169159 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" event={"ID":"2c6fc738-8526-4906-ad37-040910d43c72","Type":"ContainerStarted","Data":"5b7bc1e969120d34afdc5ec0da62b6f19cbfe8f117674ce5581e154c759c8266"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.169223 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" event={"ID":"2c6fc738-8526-4906-ad37-040910d43c72","Type":"ContainerStarted","Data":"0e23e11f34a15cde021169e1de41ca3701727297e4990c80e771adc939864902"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.169235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" event={"ID":"2c6fc738-8526-4906-ad37-040910d43c72","Type":"ContainerStarted","Data":"c3a41c8db6ec285842763f548c954d34ff567c4e47fe00697c5f54da3eb17562"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.171134 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" event={"ID":"7b287991-894d-4985-b0e0-74a46ae4a7d3","Type":"ContainerStarted","Data":"90cd29627a3734325357e18db7c1d0d7462be3b7e5fbde572599011d25b48e24"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.171169 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" event={"ID":"7b287991-894d-4985-b0e0-74a46ae4a7d3","Type":"ContainerStarted","Data":"0d3b3f4895d6c365b065a377e3e69318e33316d5562aa7598710e3a1e75bb901"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.173566 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.173666 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wrsrk" event={"ID":"177c9b4a-8758-47be-9b2e-004485032310","Type":"ContainerStarted","Data":"372540e4ce6e9983bfedb649e5562559abc3f21927ac229d21cb8edfc210c322"} Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.173682 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.673655993 +0000 UTC m=+145.662884810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.173705 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wrsrk" event={"ID":"177c9b4a-8758-47be-9b2e-004485032310","Type":"ContainerStarted","Data":"40b31da6be6e86c71858692d6b25b9468e0aa2415207cc52452fef03f3818cc1"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.173847 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wrsrk" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.176361 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.176511 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.676494416 +0000 UTC m=+145.665723213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.177314 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" event={"ID":"2df46f36-de11-4b06-86d4-7ad0e6e58737","Type":"ContainerStarted","Data":"fc4bf5bdc2df738285aa4489a205475203fa7b0178793174ec4b5ccebc2861ab"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.177349 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" event={"ID":"2df46f36-de11-4b06-86d4-7ad0e6e58737","Type":"ContainerStarted","Data":"b8d03a0334832cb8cfecb335ef055e88a2c3e4053491dbc2e779c0f130af9994"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.182090 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" event={"ID":"61c94092-2bfe-4a9a-ab80-c6268da4d42c","Type":"ContainerStarted","Data":"e6f867fa0083660175e3d60ddfcab16b3ab1a4a7770a4e56dded6c918c629ac4"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.183728 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" event={"ID":"6bc3891f-6cc8-445d-b852-6b34df2bf821","Type":"ContainerStarted","Data":"8a5abaa764b1f2b2decb3bdd7604ce2f4e2332afb08683f624a353f146750bb2"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.183770 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" event={"ID":"6bc3891f-6cc8-445d-b852-6b34df2bf821","Type":"ContainerStarted","Data":"3ebd919cfc3f97b649ebdc1412e674541562e73d2900e127f84437f7ba8e2144"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.187285 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" event={"ID":"40ccad2f-114b-43c1-8b78-d5c252dea8ff","Type":"ContainerStarted","Data":"171a7035c4ea00efb4bd32b044bf052a6b22edf9a9200e5ce856c98bd4be3644"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.197705 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" event={"ID":"1cdabf47-3a38-4c24-b2e8-40ae2f062f08","Type":"ContainerStarted","Data":"01727fbe6db3bca094bf15e7c6fce1a3be74001aa65a9a8916a0bfd5a065cb9a"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.207106 4983 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xhsfh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.207156 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.207211 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" event={"ID":"7330817e-3c60-4eaa-b546-76a64214a03f","Type":"ContainerStarted","Data":"81079fe919223fc8bec59ea3ae4b83176c1e1a88fbca871f8aeba0e17b12fef0"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.207241 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" event={"ID":"7330817e-3c60-4eaa-b546-76a64214a03f","Type":"ContainerStarted","Data":"38046382ec4135e9ca66762798065bbe6267ed5e74af08f03205a12e928a3857"} Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.211797 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fvqlm" podStartSLOduration=125.211777787 podStartE2EDuration="2m5.211777787s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.194607628 +0000 UTC m=+145.183836425" watchObservedRunningTime="2025-10-01 09:11:57.211777787 +0000 UTC m=+145.201006584" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.212095 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jzscr" podStartSLOduration=7.212088304 podStartE2EDuration="7.212088304s" podCreationTimestamp="2025-10-01 09:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.210443272 +0000 UTC m=+145.199672079" watchObservedRunningTime="2025-10-01 09:11:57.212088304 +0000 UTC m=+145.201317101" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.219445 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jclq9" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.221549 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-n26vk" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.229041 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4xzjs" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.255190 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rpg24" podStartSLOduration=125.255168883 podStartE2EDuration="2m5.255168883s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.234085655 +0000 UTC m=+145.223314462" watchObservedRunningTime="2025-10-01 09:11:57.255168883 +0000 UTC m=+145.244397680" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.279373 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.279587 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.779568775 +0000 UTC m=+145.768797572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.280129 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.281157 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.781144565 +0000 UTC m=+145.770373362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.291033 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wrsrk" podStartSLOduration=7.291014697 podStartE2EDuration="7.291014697s" podCreationTimestamp="2025-10-01 09:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.290250988 +0000 UTC m=+145.279479795" watchObservedRunningTime="2025-10-01 09:11:57.291014697 +0000 UTC m=+145.280243494" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.291415 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5stn" podStartSLOduration=125.291410797 podStartE2EDuration="2m5.291410797s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.260598241 +0000 UTC m=+145.249827038" watchObservedRunningTime="2025-10-01 09:11:57.291410797 +0000 UTC m=+145.280639594" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.306107 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qppm8" podStartSLOduration=125.306088752 podStartE2EDuration="2m5.306088752s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.305014274 +0000 UTC m=+145.294243071" watchObservedRunningTime="2025-10-01 09:11:57.306088752 +0000 UTC m=+145.295317549" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.323465 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" podStartSLOduration=125.323445904 podStartE2EDuration="2m5.323445904s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.321861644 +0000 UTC m=+145.311090441" watchObservedRunningTime="2025-10-01 09:11:57.323445904 +0000 UTC m=+145.312674711" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.374108 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5rckn" podStartSLOduration=125.374092486 podStartE2EDuration="2m5.374092486s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.373719617 +0000 UTC m=+145.362948404" watchObservedRunningTime="2025-10-01 09:11:57.374092486 +0000 UTC m=+145.363321283" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.386427 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.386589 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.886564004 +0000 UTC m=+145.875792801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.386710 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.387383 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.887373095 +0000 UTC m=+145.876601892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.414924 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvtld" podStartSLOduration=125.414907328 podStartE2EDuration="2m5.414907328s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.413001139 +0000 UTC m=+145.402229936" watchObservedRunningTime="2025-10-01 09:11:57.414907328 +0000 UTC m=+145.404136125" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.488503 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.488706 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.988672919 +0000 UTC m=+145.977901716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.488786 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.489163 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:57.989155062 +0000 UTC m=+145.978383849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.500002 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-76knj" podStartSLOduration=125.499985787 podStartE2EDuration="2m5.499985787s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.464950174 +0000 UTC m=+145.454178971" watchObservedRunningTime="2025-10-01 09:11:57.499985787 +0000 UTC m=+145.489214584" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.521514 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-4jnpd" podStartSLOduration=125.521497517 podStartE2EDuration="2m5.521497517s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.518650674 +0000 UTC m=+145.507879481" watchObservedRunningTime="2025-10-01 09:11:57.521497517 +0000 UTC m=+145.510726314" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.542153 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tgkqm" podStartSLOduration=125.542132143 podStartE2EDuration="2m5.542132143s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.534269942 +0000 UTC m=+145.523498749" watchObservedRunningTime="2025-10-01 09:11:57.542132143 +0000 UTC m=+145.531360940" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.579457 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" podStartSLOduration=125.579438854 podStartE2EDuration="2m5.579438854s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.576186642 +0000 UTC m=+145.565415469" watchObservedRunningTime="2025-10-01 09:11:57.579438854 +0000 UTC m=+145.568667651" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.591337 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.591833 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.09179743 +0000 UTC m=+146.081026227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.693088 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.693380 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.193370461 +0000 UTC m=+146.182599258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.762285 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:11:57 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:11:57 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:11:57 crc kubenswrapper[4983]: healthz check failed Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.762344 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.793906 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.794261 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.294244444 +0000 UTC m=+146.283473241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.794382 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.794713 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.294695915 +0000 UTC m=+146.283924712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.895929 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.896233 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.396207104 +0000 UTC m=+146.385435901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.896431 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.896725 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.396714018 +0000 UTC m=+146.385942815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.997155 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.997408 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.497386406 +0000 UTC m=+146.486615193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:57 crc kubenswrapper[4983]: I1001 09:11:57.997618 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:57 crc kubenswrapper[4983]: E1001 09:11:57.997950 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.49793711 +0000 UTC m=+146.487165907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.098578 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.099145 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.599131181 +0000 UTC m=+146.588359978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.199827 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.200423 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.700410955 +0000 UTC m=+146.689639752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.210874 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" event={"ID":"8501eee0-e9f1-4b16-ba33-f8ba28add8fd","Type":"ContainerStarted","Data":"b9de7e1e100bf5219e3376654e73e242eb759e3b0b26d0b31000c310be9a5dd9"} Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.220366 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.235163 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5mdqr" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.294412 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hsczv" podStartSLOduration=126.294398232 podStartE2EDuration="2m6.294398232s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:11:57.605709555 +0000 UTC m=+145.594938352" watchObservedRunningTime="2025-10-01 09:11:58.294398232 +0000 UTC m=+146.283627029" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.302525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.303762 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.80374795 +0000 UTC m=+146.792976747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.404125 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.404533 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:58.90451602 +0000 UTC m=+146.893744817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.504790 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.504910 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.004890501 +0000 UTC m=+146.994119288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.505239 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.505479 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.005471277 +0000 UTC m=+146.994700074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.605533 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.605885 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.105863817 +0000 UTC m=+147.095092624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.706574 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.707043 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.207021697 +0000 UTC m=+147.196250554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.761580 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:11:58 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:11:58 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:11:58 crc kubenswrapper[4983]: healthz check failed Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.761637 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.807172 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.807271 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.307255404 +0000 UTC m=+147.296484201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.807433 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.807699 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.307692155 +0000 UTC m=+147.296920952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.811633 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.812525 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.814073 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.822406 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.908520 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.908617 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.408602899 +0000 UTC m=+147.397831696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.908756 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:58 crc kubenswrapper[4983]: E1001 09:11:58.909004 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.40899754 +0000 UTC m=+147.398226337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.909208 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.909235 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:58 crc kubenswrapper[4983]: I1001 09:11:58.909302 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lttbm\" (UniqueName: \"kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.009868 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.010137 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.510108259 +0000 UTC m=+147.499337056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.010628 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.010734 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lttbm\" (UniqueName: \"kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.010846 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.010999 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.011446 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.011937 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.012234 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.512202971 +0000 UTC m=+147.501431758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.016572 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.017480 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.020819 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.031150 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lttbm\" (UniqueName: \"kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm\") pod \"community-operators-clr5p\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.032018 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.113614 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.114003 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.114085 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnw9x\" (UniqueName: \"kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.114144 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.114323 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.614304776 +0000 UTC m=+147.603533573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.127204 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.210027 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.211067 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.216309 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.216354 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.216410 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.216455 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnw9x\" (UniqueName: \"kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.217210 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.717193081 +0000 UTC m=+147.706421868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.217257 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.217489 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.223558 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.233682 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" event={"ID":"8501eee0-e9f1-4b16-ba33-f8ba28add8fd","Type":"ContainerStarted","Data":"70b406fc553d9a8427e8f226ecdf35d47526abd77a972e801bbff1ea46c8f471"} Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.244596 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnw9x\" (UniqueName: \"kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x\") pod \"certified-operators-2lvtf\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.317872 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.318033 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmg8z\" (UniqueName: \"kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.318067 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.318096 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.318235 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.818220748 +0000 UTC m=+147.807449545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.363969 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.403408 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.406422 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.415148 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.419906 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmg8z\" (UniqueName: \"kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.419962 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.420027 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.420063 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.422600 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.428286 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.428544 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:11:59.928530682 +0000 UTC m=+147.917759479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.451891 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmg8z\" (UniqueName: \"kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z\") pod \"community-operators-q5t7p\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.472028 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.521480 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.521662 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:12:00.021635496 +0000 UTC m=+148.010864293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.521934 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.521994 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.522059 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.522096 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqnjz\" (UniqueName: \"kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.522350 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:12:00.022337035 +0000 UTC m=+148.011565832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.531624 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.577442 4983 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.586055 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:11:59 crc kubenswrapper[4983]: W1001 09:11:59.593739 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a77c4f0_1ce1_45b4_8e9a_3c7360290e79.slice/crio-bcd6f7b86d1f2b53eb2639f04a608422a3a626aa86fbe2cfe19b83bf0baed6f2 WatchSource:0}: Error finding container bcd6f7b86d1f2b53eb2639f04a608422a3a626aa86fbe2cfe19b83bf0baed6f2: Status 404 returned error can't find the container with id bcd6f7b86d1f2b53eb2639f04a608422a3a626aa86fbe2cfe19b83bf0baed6f2 Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.623724 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.623890 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.623947 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.623988 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.624008 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.624024 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqnjz\" (UniqueName: \"kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.624048 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.624074 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.624548 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.624982 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:12:00.124950222 +0000 UTC m=+148.114179019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.625389 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.625469 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.629403 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.633217 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.640588 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.649727 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqnjz\" (UniqueName: \"kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz\") pod \"certified-operators-22tvh\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.655376 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.669003 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.725439 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.726017 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:12:00.22600649 +0000 UTC m=+148.215235287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2pr5s" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.757350 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.762475 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.766042 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:11:59 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:11:59 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:11:59 crc kubenswrapper[4983]: healthz check failed Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.766083 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.826297 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:11:59 crc kubenswrapper[4983]: E1001 09:11:59.826645 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:12:00.326630417 +0000 UTC m=+148.315859204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.888424 4983 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-01T09:11:59.577466721Z","Handler":null,"Name":""} Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.893275 4983 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.893304 4983 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.930971 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.933102 4983 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.933135 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.940056 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:11:59 crc kubenswrapper[4983]: I1001 09:11:59.978636 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2pr5s\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.033085 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.048476 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.090449 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.120517 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.251088 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" event={"ID":"8501eee0-e9f1-4b16-ba33-f8ba28add8fd","Type":"ContainerStarted","Data":"10928a0ef111f4f0d203f5b4d32c3ab6166b7e3448139e8ce949091908bca9cb"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.251409 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" event={"ID":"8501eee0-e9f1-4b16-ba33-f8ba28add8fd","Type":"ContainerStarted","Data":"908f698fc39e0fd0a71a0c40b1b2599cb8b4ac4737b1e72e9d9c620c8b2eb7bc"} Oct 01 09:12:00 crc kubenswrapper[4983]: W1001 09:12:00.257931 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-c7dca70ec2a20cae07b79655530d098301438653a2f15bfc77776015b28a704a WatchSource:0}: Error finding container c7dca70ec2a20cae07b79655530d098301438653a2f15bfc77776015b28a704a: Status 404 returned error can't find the container with id c7dca70ec2a20cae07b79655530d098301438653a2f15bfc77776015b28a704a Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.258338 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d1b419699926d99be7ef0c0675e40e7ad3c548e7391fcd93e10aefab8c20dab2"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.265891 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-92fqg" podStartSLOduration=10.265877291 podStartE2EDuration="10.265877291s" podCreationTimestamp="2025-10-01 09:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:00.265038359 +0000 UTC m=+148.254267156" watchObservedRunningTime="2025-10-01 09:12:00.265877291 +0000 UTC m=+148.255106088" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.270677 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"13f82efb633d6cbe780d7cd43f90c3cf7e371d4388994b007584928b837f4813"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.272906 4983 generic.go:334] "Generic (PLEG): container finished" podID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerID="a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56" exitCode=0 Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.272988 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerDied","Data":"a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.273015 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerStarted","Data":"4efa6b5685bd4815fc5baf0bc2f59a384a75e9667820bb77e614363860246fd5"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.276104 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.299850 4983 generic.go:334] "Generic (PLEG): container finished" podID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerID="c1c9c82291b10d4386f6367c6920f84424e6e8663b2f3ff54853f8aed9cd36ac" exitCode=0 Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.299919 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerDied","Data":"c1c9c82291b10d4386f6367c6920f84424e6e8663b2f3ff54853f8aed9cd36ac"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.299949 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerStarted","Data":"bcd6f7b86d1f2b53eb2639f04a608422a3a626aa86fbe2cfe19b83bf0baed6f2"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.304535 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerStarted","Data":"1ad97990912ab7e4367bc322ccd066ab4b1c20459578850767740e442b3274c0"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.310324 4983 generic.go:334] "Generic (PLEG): container finished" podID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerID="49e3b784eb98dc25dceb501f9297174669112e978833d4271c62efe3b90bc4c3" exitCode=0 Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.310406 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerDied","Data":"49e3b784eb98dc25dceb501f9297174669112e978833d4271c62efe3b90bc4c3"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.310454 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerStarted","Data":"8aee3124f860a5c5e3c3761350aac8c7aa0ea883ed0cbc51a38728069bec1684"} Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.324747 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:12:00 crc kubenswrapper[4983]: W1001 09:12:00.344440 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbde163a6_2b23_4f0f_873b_a37d2213d232.slice/crio-87cb0ccdb84c5232ccd43c35871c8af02ee61d818379f598de49bfb468bb92b5 WatchSource:0}: Error finding container 87cb0ccdb84c5232ccd43c35871c8af02ee61d818379f598de49bfb468bb92b5: Status 404 returned error can't find the container with id 87cb0ccdb84c5232ccd43c35871c8af02ee61d818379f598de49bfb468bb92b5 Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.721243 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.761664 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:12:00 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:12:00 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:12:00 crc kubenswrapper[4983]: healthz check failed Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.761723 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.810408 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.812017 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.814212 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.820087 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.944503 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.944626 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:00 crc kubenswrapper[4983]: I1001 09:12:00.944655 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qszfz\" (UniqueName: \"kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.046316 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.046361 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qszfz\" (UniqueName: \"kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.046444 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.046829 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.047230 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.084507 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qszfz\" (UniqueName: \"kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz\") pod \"redhat-marketplace-ptxwd\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.128648 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.206942 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.209182 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.224388 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.342563 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"84d920c2748a705e044f2801cfb19b6d74209d911becec91f44cac790ad49c82"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.349682 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.349710 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.349744 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz5hp\" (UniqueName: \"kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.352688 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a3493fb076e3c5eb050459ff9acf8f4a63ac4e31bf313c20e65976c09cbb7097"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.352746 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c7dca70ec2a20cae07b79655530d098301438653a2f15bfc77776015b28a704a"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.372703 4983 generic.go:334] "Generic (PLEG): container finished" podID="40ccad2f-114b-43c1-8b78-d5c252dea8ff" containerID="171a7035c4ea00efb4bd32b044bf052a6b22edf9a9200e5ce856c98bd4be3644" exitCode=0 Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.372868 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" event={"ID":"40ccad2f-114b-43c1-8b78-d5c252dea8ff","Type":"ContainerDied","Data":"171a7035c4ea00efb4bd32b044bf052a6b22edf9a9200e5ce856c98bd4be3644"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.442905 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e37e7a3996d4746c44265f6ddd35b2484f69f7367ca6e74362621db1a04332e6"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.443833 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.453339 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz5hp\" (UniqueName: \"kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.454758 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.454792 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.455473 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.458680 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.484994 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.489526 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" event={"ID":"bde163a6-2b23-4f0f-873b-a37d2213d232","Type":"ContainerStarted","Data":"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.489555 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" event={"ID":"bde163a6-2b23-4f0f-873b-a37d2213d232","Type":"ContainerStarted","Data":"87cb0ccdb84c5232ccd43c35871c8af02ee61d818379f598de49bfb468bb92b5"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.489836 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.501526 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz5hp\" (UniqueName: \"kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp\") pod \"redhat-marketplace-l5882\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.511529 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" podStartSLOduration=129.511513785 podStartE2EDuration="2m9.511513785s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:01.511392522 +0000 UTC m=+149.500621349" watchObservedRunningTime="2025-10-01 09:12:01.511513785 +0000 UTC m=+149.500742582" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.513500 4983 generic.go:334] "Generic (PLEG): container finished" podID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerID="2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb" exitCode=0 Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.513629 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerDied","Data":"2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb"} Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.533568 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.762348 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:12:01 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:12:01 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:12:01 crc kubenswrapper[4983]: healthz check failed Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.762675 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.780826 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.786237 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bkjc6" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.804444 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.920896 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.921248 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.924770 4983 patch_prober.go:28] interesting pod/console-f9d7485db-tf547 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.924837 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-tf547" podUID="af78a493-d342-4762-87e6-fdfdda346927" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.951151 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-jjb8k container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.951245 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jjb8k" podUID="46bfb1e9-ebaf-4e47-9983-71e6d6a01154" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.951197 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-jjb8k container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.951322 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jjb8k" podUID="46bfb1e9-ebaf-4e47-9983-71e6d6a01154" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.962913 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:12:01 crc kubenswrapper[4983]: I1001 09:12:01.962948 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.021451 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.022740 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.025147 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.030735 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.031654 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.034302 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.034504 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.065163 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.067845 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.167219 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.167278 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprlf\" (UniqueName: \"kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.167302 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.167330 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.167360 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.268869 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.268946 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprlf\" (UniqueName: \"kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.268974 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.269013 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.269048 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.269170 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.270704 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.271005 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.290836 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.309637 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprlf\" (UniqueName: \"kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf\") pod \"redhat-operators-cdcrl\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.348131 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.366751 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.426338 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.427834 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.449988 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.572352 4983 generic.go:334] "Generic (PLEG): container finished" podID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerID="a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6" exitCode=0 Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.572477 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerDied","Data":"a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6"} Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.572678 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerStarted","Data":"7a39c423a140e6e3124b0ded323f50cca882a8532fb6fc6e55789a22b2fe299b"} Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.575177 4983 generic.go:334] "Generic (PLEG): container finished" podID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerID="413d4fd2ca9d96721f056e97de7901459899b3e6198468266d6c430dd771a673" exitCode=0 Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.575392 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerDied","Data":"413d4fd2ca9d96721f056e97de7901459899b3e6198468266d6c430dd771a673"} Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.575465 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerStarted","Data":"0f09f1295ecf84fc7944ca88b0f77429eccc720724af2d80a4e04bedf5b50bcc"} Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.581546 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.581848 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.581871 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mxml\" (UniqueName: \"kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.682683 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.682746 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.682772 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mxml\" (UniqueName: \"kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.683439 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.684604 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.685163 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.704917 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mxml\" (UniqueName: \"kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml\") pod \"redhat-operators-8z7nl\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.750655 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.762714 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:12:02 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Oct 01 09:12:02 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:12:02 crc kubenswrapper[4983]: healthz check failed Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.762777 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:12:02 crc kubenswrapper[4983]: I1001 09:12:02.905860 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.012315 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume\") pod \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.012404 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume\") pod \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.012471 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwmbn\" (UniqueName: \"kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn\") pod \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\" (UID: \"40ccad2f-114b-43c1-8b78-d5c252dea8ff\") " Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.013509 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume" (OuterVolumeSpecName: "config-volume") pod "40ccad2f-114b-43c1-8b78-d5c252dea8ff" (UID: "40ccad2f-114b-43c1-8b78-d5c252dea8ff"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.019674 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ccad2f-114b-43c1-8b78-d5c252dea8ff-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.026190 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn" (OuterVolumeSpecName: "kube-api-access-dwmbn") pod "40ccad2f-114b-43c1-8b78-d5c252dea8ff" (UID: "40ccad2f-114b-43c1-8b78-d5c252dea8ff"). InnerVolumeSpecName "kube-api-access-dwmbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.026258 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40ccad2f-114b-43c1-8b78-d5c252dea8ff" (UID: "40ccad2f-114b-43c1-8b78-d5c252dea8ff"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.056190 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:12:03 crc kubenswrapper[4983]: W1001 09:12:03.120196 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod75061baf_2309_4dcc_be5b_eb39bff7dfea.slice/crio-ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6 WatchSource:0}: Error finding container ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6: Status 404 returned error can't find the container with id ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6 Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.120666 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwmbn\" (UniqueName: \"kubernetes.io/projected/40ccad2f-114b-43c1-8b78-d5c252dea8ff-kube-api-access-dwmbn\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.120697 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ccad2f-114b-43c1-8b78-d5c252dea8ff-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.252103 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.589910 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75061baf-2309-4dcc-be5b-eb39bff7dfea","Type":"ContainerStarted","Data":"ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6"} Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.594970 4983 generic.go:334] "Generic (PLEG): container finished" podID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerID="354db233e23a8c6c5a9c780ab22d3b7c037eb4f91dc1de438865ebbbbf39b00a" exitCode=0 Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.595053 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerDied","Data":"354db233e23a8c6c5a9c780ab22d3b7c037eb4f91dc1de438865ebbbbf39b00a"} Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.595105 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerStarted","Data":"bc75a305119b6ae84a2e3bd95bd924f5956d3bfa0a91e3af9cc29f220f4cd058"} Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.600665 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.600677 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-vlk57" event={"ID":"40ccad2f-114b-43c1-8b78-d5c252dea8ff","Type":"ContainerDied","Data":"960fb05200c6b763c18e99a08c253bf554452da41fbd01ed19b4c60b8cbd2f03"} Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.601126 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="960fb05200c6b763c18e99a08c253bf554452da41fbd01ed19b4c60b8cbd2f03" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.607998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerStarted","Data":"7de5f5b6626dc60e5cef05990e4f13e8a766f0dc9d3720dd29d5605ea7c6d8b9"} Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.758738 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.761397 4983 patch_prober.go:28] interesting pod/router-default-5444994796-5wzgd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:12:03 crc kubenswrapper[4983]: [+]has-synced ok Oct 01 09:12:03 crc kubenswrapper[4983]: [+]process-running ok Oct 01 09:12:03 crc kubenswrapper[4983]: healthz check failed Oct 01 09:12:03 crc kubenswrapper[4983]: I1001 09:12:03.761433 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5wzgd" podUID="11a442c1-aca4-467a-bbb7-5ccea25def0b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.618512 4983 generic.go:334] "Generic (PLEG): container finished" podID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerID="f3ca02920b77761d1a8c349772b4f62819c080929708779a36408a86ada7ce5c" exitCode=0 Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.618751 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerDied","Data":"f3ca02920b77761d1a8c349772b4f62819c080929708779a36408a86ada7ce5c"} Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.627955 4983 generic.go:334] "Generic (PLEG): container finished" podID="75061baf-2309-4dcc-be5b-eb39bff7dfea" containerID="e5b0c5b014c6c67bcde27cbb34b9f65c167ff6f00eeda50b525c7df790c0b715" exitCode=0 Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.628018 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75061baf-2309-4dcc-be5b-eb39bff7dfea","Type":"ContainerDied","Data":"e5b0c5b014c6c67bcde27cbb34b9f65c167ff6f00eeda50b525c7df790c0b715"} Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.761340 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:12:04 crc kubenswrapper[4983]: I1001 09:12:04.763408 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5wzgd" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.643006 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:12:05 crc kubenswrapper[4983]: E1001 09:12:05.643496 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ccad2f-114b-43c1-8b78-d5c252dea8ff" containerName="collect-profiles" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.643507 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ccad2f-114b-43c1-8b78-d5c252dea8ff" containerName="collect-profiles" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.643643 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ccad2f-114b-43c1-8b78-d5c252dea8ff" containerName="collect-profiles" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.644164 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.651687 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.652338 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.685918 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.686005 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.708266 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.790149 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.790283 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.790607 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.814381 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:05 crc kubenswrapper[4983]: I1001 09:12:05.974267 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:08 crc kubenswrapper[4983]: I1001 09:12:08.566766 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wrsrk" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.646232 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.745521 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75061baf-2309-4dcc-be5b-eb39bff7dfea","Type":"ContainerDied","Data":"ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6"} Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.745576 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba162f6549b30a8803eda891a02152479ce5ed19769158795cf5b49c2a7e60e6" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.745673 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.787083 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir\") pod \"75061baf-2309-4dcc-be5b-eb39bff7dfea\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.787320 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access\") pod \"75061baf-2309-4dcc-be5b-eb39bff7dfea\" (UID: \"75061baf-2309-4dcc-be5b-eb39bff7dfea\") " Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.787327 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "75061baf-2309-4dcc-be5b-eb39bff7dfea" (UID: "75061baf-2309-4dcc-be5b-eb39bff7dfea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.789164 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75061baf-2309-4dcc-be5b-eb39bff7dfea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.794737 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "75061baf-2309-4dcc-be5b-eb39bff7dfea" (UID: "75061baf-2309-4dcc-be5b-eb39bff7dfea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:10 crc kubenswrapper[4983]: I1001 09:12:10.890590 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75061baf-2309-4dcc-be5b-eb39bff7dfea-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:11 crc kubenswrapper[4983]: I1001 09:12:11.921663 4983 patch_prober.go:28] interesting pod/console-f9d7485db-tf547 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 01 09:12:11 crc kubenswrapper[4983]: I1001 09:12:11.921722 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-tf547" podUID="af78a493-d342-4762-87e6-fdfdda346927" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 01 09:12:11 crc kubenswrapper[4983]: I1001 09:12:11.957514 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-jjb8k" Oct 01 09:12:14 crc kubenswrapper[4983]: I1001 09:12:14.238461 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:12:14 crc kubenswrapper[4983]: I1001 09:12:14.268745 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a2350dc-e1ed-4b83-8fbd-6b8321f34895-metrics-certs\") pod \"network-metrics-daemon-pvhnl\" (UID: \"7a2350dc-e1ed-4b83-8fbd-6b8321f34895\") " pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:12:14 crc kubenswrapper[4983]: I1001 09:12:14.535378 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvhnl" Oct 01 09:12:20 crc kubenswrapper[4983]: I1001 09:12:20.099459 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:12:21 crc kubenswrapper[4983]: I1001 09:12:21.929190 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:12:21 crc kubenswrapper[4983]: I1001 09:12:21.935890 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-tf547" Oct 01 09:12:22 crc kubenswrapper[4983]: E1001 09:12:22.349873 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 09:12:22 crc kubenswrapper[4983]: E1001 09:12:22.350443 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqnjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-22tvh_openshift-marketplace(2e6078c0-b5c1-488d-9ed3-ff31fb87dd49): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:22 crc kubenswrapper[4983]: E1001 09:12:22.351907 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-22tvh" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" Oct 01 09:12:23 crc kubenswrapper[4983]: E1001 09:12:23.618052 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-22tvh" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" Oct 01 09:12:23 crc kubenswrapper[4983]: E1001 09:12:23.721135 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 09:12:23 crc kubenswrapper[4983]: E1001 09:12:23.721467 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qszfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ptxwd_openshift-marketplace(ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:23 crc kubenswrapper[4983]: E1001 09:12:23.722733 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ptxwd" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.844995 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ptxwd" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.930406 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.930764 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nz5hp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l5882_openshift-marketplace(2f86c554-fecb-48e6-b257-8f398bdbc97d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.932568 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l5882" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.951445 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.951629 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nmg8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-q5t7p_openshift-marketplace(48f780d0-874f-4d5e-b48e-79abb6cf0e28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.953858 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-q5t7p" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.966438 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.966789 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nnw9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2lvtf_openshift-marketplace(8a77c4f0-1ce1-45b4-8e9a-3c7360290e79): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.968418 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2lvtf" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.974431 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.974553 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lttbm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-clr5p_openshift-marketplace(b9f5a446-fb85-404e-9200-f38d58f5442f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:12:24 crc kubenswrapper[4983]: E1001 09:12:24.975901 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-clr5p" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" Oct 01 09:12:25 crc kubenswrapper[4983]: I1001 09:12:25.008492 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:12:27 crc kubenswrapper[4983]: E1001 09:12:27.605014 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-clr5p" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" Oct 01 09:12:27 crc kubenswrapper[4983]: E1001 09:12:27.605367 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l5882" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" Oct 01 09:12:27 crc kubenswrapper[4983]: E1001 09:12:27.605417 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2lvtf" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" Oct 01 09:12:27 crc kubenswrapper[4983]: E1001 09:12:27.605463 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-q5t7p" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" Oct 01 09:12:27 crc kubenswrapper[4983]: I1001 09:12:27.837695 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53fd5c81-7bb2-4411-8d0c-9a0ad13da649","Type":"ContainerStarted","Data":"14554e4f7470214fd8bc89918e6baeb6ae65573debe39abd11510f102870f278"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.003612 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pvhnl"] Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.858874 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" event={"ID":"7a2350dc-e1ed-4b83-8fbd-6b8321f34895","Type":"ContainerStarted","Data":"7db9f20f08ecf30cb87147b6d23999a315c69669d93c8ec6edf6ae46d66a0ba0"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.859235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" event={"ID":"7a2350dc-e1ed-4b83-8fbd-6b8321f34895","Type":"ContainerStarted","Data":"2500a1d1f934498da73c287e15fba8f80061e627f9ed795e30f8843c3f475e5c"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.859266 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvhnl" event={"ID":"7a2350dc-e1ed-4b83-8fbd-6b8321f34895","Type":"ContainerStarted","Data":"d793e05a37e229f001df4c7d6ee7a103712d7a30d3f3502b738115ef5f5116a6"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.863329 4983 generic.go:334] "Generic (PLEG): container finished" podID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerID="ee659ae610366ae39530b905c08421420d6c2a34ef7a73d3e2cf29cf892c58fd" exitCode=0 Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.863460 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerDied","Data":"ee659ae610366ae39530b905c08421420d6c2a34ef7a73d3e2cf29cf892c58fd"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.868181 4983 generic.go:334] "Generic (PLEG): container finished" podID="53fd5c81-7bb2-4411-8d0c-9a0ad13da649" containerID="7c47ffaef194ecbde0d0f077cd9586b3f2733303791939ebebb209466f413923" exitCode=0 Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.868296 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53fd5c81-7bb2-4411-8d0c-9a0ad13da649","Type":"ContainerDied","Data":"7c47ffaef194ecbde0d0f077cd9586b3f2733303791939ebebb209466f413923"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.873463 4983 generic.go:334] "Generic (PLEG): container finished" podID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerID="901a7eec258c9b0cb9ffab21dca6544cf3c2fa5bb29a503fdcf6d21c6848d250" exitCode=0 Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.873495 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerDied","Data":"901a7eec258c9b0cb9ffab21dca6544cf3c2fa5bb29a503fdcf6d21c6848d250"} Oct 01 09:12:28 crc kubenswrapper[4983]: I1001 09:12:28.883221 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-pvhnl" podStartSLOduration=156.883208572 podStartE2EDuration="2m36.883208572s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:28.882052953 +0000 UTC m=+176.871281760" watchObservedRunningTime="2025-10-01 09:12:28.883208572 +0000 UTC m=+176.872437379" Oct 01 09:12:29 crc kubenswrapper[4983]: I1001 09:12:29.883572 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerStarted","Data":"816976e74a03122aa7f3e14d788c9d592d3318656e7771a2dcb742716afcc077"} Oct 01 09:12:29 crc kubenswrapper[4983]: I1001 09:12:29.885713 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerStarted","Data":"1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1"} Oct 01 09:12:29 crc kubenswrapper[4983]: I1001 09:12:29.913053 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8z7nl" podStartSLOduration=3.242437634 podStartE2EDuration="27.913030671s" podCreationTimestamp="2025-10-01 09:12:02 +0000 UTC" firstStartedPulling="2025-10-01 09:12:04.622620324 +0000 UTC m=+152.611849121" lastFinishedPulling="2025-10-01 09:12:29.293213341 +0000 UTC m=+177.282442158" observedRunningTime="2025-10-01 09:12:29.906842774 +0000 UTC m=+177.896071611" watchObservedRunningTime="2025-10-01 09:12:29.913030671 +0000 UTC m=+177.902259478" Oct 01 09:12:29 crc kubenswrapper[4983]: I1001 09:12:29.930647 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cdcrl" podStartSLOduration=3.157233303 podStartE2EDuration="28.93062835s" podCreationTimestamp="2025-10-01 09:12:01 +0000 UTC" firstStartedPulling="2025-10-01 09:12:03.597500855 +0000 UTC m=+151.586729652" lastFinishedPulling="2025-10-01 09:12:29.370895862 +0000 UTC m=+177.360124699" observedRunningTime="2025-10-01 09:12:29.92669522 +0000 UTC m=+177.915924047" watchObservedRunningTime="2025-10-01 09:12:29.93062835 +0000 UTC m=+177.919857147" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.120078 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.245468 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access\") pod \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.245637 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir\") pod \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\" (UID: \"53fd5c81-7bb2-4411-8d0c-9a0ad13da649\") " Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.245751 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "53fd5c81-7bb2-4411-8d0c-9a0ad13da649" (UID: "53fd5c81-7bb2-4411-8d0c-9a0ad13da649"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.246148 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.257014 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "53fd5c81-7bb2-4411-8d0c-9a0ad13da649" (UID: "53fd5c81-7bb2-4411-8d0c-9a0ad13da649"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.347686 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53fd5c81-7bb2-4411-8d0c-9a0ad13da649-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.891689 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53fd5c81-7bb2-4411-8d0c-9a0ad13da649","Type":"ContainerDied","Data":"14554e4f7470214fd8bc89918e6baeb6ae65573debe39abd11510f102870f278"} Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.891758 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14554e4f7470214fd8bc89918e6baeb6ae65573debe39abd11510f102870f278" Oct 01 09:12:30 crc kubenswrapper[4983]: I1001 09:12:30.891880 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:12:31 crc kubenswrapper[4983]: I1001 09:12:31.961128 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:12:31 crc kubenswrapper[4983]: I1001 09:12:31.961212 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:12:32 crc kubenswrapper[4983]: I1001 09:12:32.348333 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:32 crc kubenswrapper[4983]: I1001 09:12:32.348390 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:32 crc kubenswrapper[4983]: I1001 09:12:32.752071 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:32 crc kubenswrapper[4983]: I1001 09:12:32.752423 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:33 crc kubenswrapper[4983]: I1001 09:12:33.186545 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b2j8m" Oct 01 09:12:33 crc kubenswrapper[4983]: I1001 09:12:33.500640 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cdcrl" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" probeResult="failure" output=< Oct 01 09:12:33 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:12:33 crc kubenswrapper[4983]: > Oct 01 09:12:33 crc kubenswrapper[4983]: I1001 09:12:33.796507 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8z7nl" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="registry-server" probeResult="failure" output=< Oct 01 09:12:33 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:12:33 crc kubenswrapper[4983]: > Oct 01 09:12:39 crc kubenswrapper[4983]: I1001 09:12:39.760786 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:12:39 crc kubenswrapper[4983]: I1001 09:12:39.951903 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerStarted","Data":"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716"} Oct 01 09:12:40 crc kubenswrapper[4983]: I1001 09:12:40.969064 4983 generic.go:334] "Generic (PLEG): container finished" podID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerID="58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716" exitCode=0 Oct 01 09:12:40 crc kubenswrapper[4983]: I1001 09:12:40.969471 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerDied","Data":"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716"} Oct 01 09:12:42 crc kubenswrapper[4983]: I1001 09:12:42.413017 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:42 crc kubenswrapper[4983]: I1001 09:12:42.462222 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:12:42 crc kubenswrapper[4983]: I1001 09:12:42.790542 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:42 crc kubenswrapper[4983]: I1001 09:12:42.875049 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:44 crc kubenswrapper[4983]: I1001 09:12:44.959863 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:44 crc kubenswrapper[4983]: I1001 09:12:44.960079 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8z7nl" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="registry-server" containerID="cri-o://816976e74a03122aa7f3e14d788c9d592d3318656e7771a2dcb742716afcc077" gracePeriod=2 Oct 01 09:12:45 crc kubenswrapper[4983]: I1001 09:12:45.997413 4983 generic.go:334] "Generic (PLEG): container finished" podID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerID="816976e74a03122aa7f3e14d788c9d592d3318656e7771a2dcb742716afcc077" exitCode=0 Oct 01 09:12:45 crc kubenswrapper[4983]: I1001 09:12:45.997490 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerDied","Data":"816976e74a03122aa7f3e14d788c9d592d3318656e7771a2dcb742716afcc077"} Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.811761 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.881258 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content\") pod \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.881616 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities\") pod \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.881695 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mxml\" (UniqueName: \"kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml\") pod \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\" (UID: \"bf90392d-40b6-400b-941b-0fcd6bc8b30b\") " Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.882604 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities" (OuterVolumeSpecName: "utilities") pod "bf90392d-40b6-400b-941b-0fcd6bc8b30b" (UID: "bf90392d-40b6-400b-941b-0fcd6bc8b30b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.889079 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml" (OuterVolumeSpecName: "kube-api-access-9mxml") pod "bf90392d-40b6-400b-941b-0fcd6bc8b30b" (UID: "bf90392d-40b6-400b-941b-0fcd6bc8b30b"). InnerVolumeSpecName "kube-api-access-9mxml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.974154 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf90392d-40b6-400b-941b-0fcd6bc8b30b" (UID: "bf90392d-40b6-400b-941b-0fcd6bc8b30b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.983579 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.983643 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mxml\" (UniqueName: \"kubernetes.io/projected/bf90392d-40b6-400b-941b-0fcd6bc8b30b-kube-api-access-9mxml\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:46 crc kubenswrapper[4983]: I1001 09:12:46.983680 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf90392d-40b6-400b-941b-0fcd6bc8b30b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.005398 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z7nl" event={"ID":"bf90392d-40b6-400b-941b-0fcd6bc8b30b","Type":"ContainerDied","Data":"7de5f5b6626dc60e5cef05990e4f13e8a766f0dc9d3720dd29d5605ea7c6d8b9"} Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.005446 4983 scope.go:117] "RemoveContainer" containerID="816976e74a03122aa7f3e14d788c9d592d3318656e7771a2dcb742716afcc077" Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.005605 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z7nl" Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.009873 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerStarted","Data":"bafae28a91eb2faa1e632e679bc654ca53542500ac41c5b863b862df6c296fae"} Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.014684 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerStarted","Data":"2cd655e71335624596cb1aa2dd7b70ca4fd9eb5a5623394ee30232b765c834ed"} Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.016347 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerStarted","Data":"0287ffe159d3e7502e7b58f4c77872edb4d82d1103a05cf76944e16f91a6f8ba"} Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.045735 4983 scope.go:117] "RemoveContainer" containerID="901a7eec258c9b0cb9ffab21dca6544cf3c2fa5bb29a503fdcf6d21c6848d250" Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.081493 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.088216 4983 scope.go:117] "RemoveContainer" containerID="f3ca02920b77761d1a8c349772b4f62819c080929708779a36408a86ada7ce5c" Oct 01 09:12:47 crc kubenswrapper[4983]: I1001 09:12:47.096103 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8z7nl"] Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.024969 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerStarted","Data":"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.026429 4983 generic.go:334] "Generic (PLEG): container finished" podID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerID="02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185" exitCode=0 Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.026492 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerDied","Data":"02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.029034 4983 generic.go:334] "Generic (PLEG): container finished" podID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerID="0287ffe159d3e7502e7b58f4c77872edb4d82d1103a05cf76944e16f91a6f8ba" exitCode=0 Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.029085 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerDied","Data":"0287ffe159d3e7502e7b58f4c77872edb4d82d1103a05cf76944e16f91a6f8ba"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.032896 4983 generic.go:334] "Generic (PLEG): container finished" podID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerID="ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a" exitCode=0 Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.033005 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerDied","Data":"ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.036786 4983 generic.go:334] "Generic (PLEG): container finished" podID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerID="bafae28a91eb2faa1e632e679bc654ca53542500ac41c5b863b862df6c296fae" exitCode=0 Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.036885 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerDied","Data":"bafae28a91eb2faa1e632e679bc654ca53542500ac41c5b863b862df6c296fae"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.039796 4983 generic.go:334] "Generic (PLEG): container finished" podID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerID="2cd655e71335624596cb1aa2dd7b70ca4fd9eb5a5623394ee30232b765c834ed" exitCode=0 Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.039865 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerDied","Data":"2cd655e71335624596cb1aa2dd7b70ca4fd9eb5a5623394ee30232b765c834ed"} Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.053165 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-22tvh" podStartSLOduration=3.674735677 podStartE2EDuration="49.053143836s" podCreationTimestamp="2025-10-01 09:11:59 +0000 UTC" firstStartedPulling="2025-10-01 09:12:01.515513547 +0000 UTC m=+149.504742344" lastFinishedPulling="2025-10-01 09:12:46.893921696 +0000 UTC m=+194.883150503" observedRunningTime="2025-10-01 09:12:48.051482887 +0000 UTC m=+196.040711704" watchObservedRunningTime="2025-10-01 09:12:48.053143836 +0000 UTC m=+196.042372633" Oct 01 09:12:48 crc kubenswrapper[4983]: I1001 09:12:48.722261 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" path="/var/lib/kubelet/pods/bf90392d-40b6-400b-941b-0fcd6bc8b30b/volumes" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.047848 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerStarted","Data":"1f8637028c7e6da5f8e78214336d4d3a9fd1a74bc728665818df0d1b8cbfc53c"} Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.050290 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerStarted","Data":"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4"} Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.052305 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerStarted","Data":"1cf101d20ea9972ece00fe7940d38c67aa046aaa835a77fdf3f2434ec8931154"} Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.054729 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerStarted","Data":"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00"} Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.057354 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerStarted","Data":"e9565826580afc48bc82ead6595de1dd0da52ab850e186bd35db54668b827706"} Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.068981 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2lvtf" podStartSLOduration=1.828378117 podStartE2EDuration="50.068962224s" podCreationTimestamp="2025-10-01 09:11:59 +0000 UTC" firstStartedPulling="2025-10-01 09:12:00.301183722 +0000 UTC m=+148.290412519" lastFinishedPulling="2025-10-01 09:12:48.541767829 +0000 UTC m=+196.530996626" observedRunningTime="2025-10-01 09:12:49.066451199 +0000 UTC m=+197.055680006" watchObservedRunningTime="2025-10-01 09:12:49.068962224 +0000 UTC m=+197.058191021" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.084317 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l5882" podStartSLOduration=2.221183966 podStartE2EDuration="48.084295852s" podCreationTimestamp="2025-10-01 09:12:01 +0000 UTC" firstStartedPulling="2025-10-01 09:12:02.573843573 +0000 UTC m=+150.563072370" lastFinishedPulling="2025-10-01 09:12:48.436955459 +0000 UTC m=+196.426184256" observedRunningTime="2025-10-01 09:12:49.081566381 +0000 UTC m=+197.070795198" watchObservedRunningTime="2025-10-01 09:12:49.084295852 +0000 UTC m=+197.073524649" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.121985 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ptxwd" podStartSLOduration=3.042319259 podStartE2EDuration="49.121963876s" podCreationTimestamp="2025-10-01 09:12:00 +0000 UTC" firstStartedPulling="2025-10-01 09:12:02.57722835 +0000 UTC m=+150.566457137" lastFinishedPulling="2025-10-01 09:12:48.656872957 +0000 UTC m=+196.646101754" observedRunningTime="2025-10-01 09:12:49.102436444 +0000 UTC m=+197.091665241" watchObservedRunningTime="2025-10-01 09:12:49.121963876 +0000 UTC m=+197.111192673" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.124281 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-clr5p" podStartSLOduration=2.91562771 podStartE2EDuration="51.124269736s" podCreationTimestamp="2025-10-01 09:11:58 +0000 UTC" firstStartedPulling="2025-10-01 09:12:00.313233089 +0000 UTC m=+148.302461886" lastFinishedPulling="2025-10-01 09:12:48.521875115 +0000 UTC m=+196.511103912" observedRunningTime="2025-10-01 09:12:49.121563465 +0000 UTC m=+197.110792262" watchObservedRunningTime="2025-10-01 09:12:49.124269736 +0000 UTC m=+197.113498533" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.128423 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.128476 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.139545 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q5t7p" podStartSLOduration=1.775378785 podStartE2EDuration="50.139522221s" podCreationTimestamp="2025-10-01 09:11:59 +0000 UTC" firstStartedPulling="2025-10-01 09:12:00.275735093 +0000 UTC m=+148.264963890" lastFinishedPulling="2025-10-01 09:12:48.639878509 +0000 UTC m=+196.629107326" observedRunningTime="2025-10-01 09:12:49.139352106 +0000 UTC m=+197.128580913" watchObservedRunningTime="2025-10-01 09:12:49.139522221 +0000 UTC m=+197.128751018" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.364802 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.364892 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.531982 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.532178 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.757514 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.757578 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:12:49 crc kubenswrapper[4983]: I1001 09:12:49.800031 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:12:50 crc kubenswrapper[4983]: I1001 09:12:50.177294 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-clr5p" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="registry-server" probeResult="failure" output=< Oct 01 09:12:50 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:12:50 crc kubenswrapper[4983]: > Oct 01 09:12:50 crc kubenswrapper[4983]: I1001 09:12:50.413604 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2lvtf" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="registry-server" probeResult="failure" output=< Oct 01 09:12:50 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:12:50 crc kubenswrapper[4983]: > Oct 01 09:12:50 crc kubenswrapper[4983]: I1001 09:12:50.576278 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-q5t7p" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="registry-server" probeResult="failure" output=< Oct 01 09:12:50 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:12:50 crc kubenswrapper[4983]: > Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.130021 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.130110 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.179511 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.534970 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.535034 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:51 crc kubenswrapper[4983]: I1001 09:12:51.572068 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:12:53 crc kubenswrapper[4983]: I1001 09:12:53.139162 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.174366 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.209386 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.405562 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.442332 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.577590 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.619346 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:12:59 crc kubenswrapper[4983]: I1001 09:12:59.801117 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.159830 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.160098 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q5t7p" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="registry-server" containerID="cri-o://2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4" gracePeriod=2 Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.571755 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.758246 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.758475 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-22tvh" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="registry-server" containerID="cri-o://d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6" gracePeriod=2 Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.960634 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.960925 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.960961 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.961432 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:13:01 crc kubenswrapper[4983]: I1001 09:13:01.961512 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198" gracePeriod=600 Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.046512 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.092051 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.133151 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198" exitCode=0 Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.133225 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198"} Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.140147 4983 generic.go:334] "Generic (PLEG): container finished" podID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerID="d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6" exitCode=0 Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.140189 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22tvh" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.140233 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerDied","Data":"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6"} Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.140269 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22tvh" event={"ID":"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49","Type":"ContainerDied","Data":"1ad97990912ab7e4367bc322ccd066ab4b1c20459578850767740e442b3274c0"} Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.140287 4983 scope.go:117] "RemoveContainer" containerID="d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.146209 4983 generic.go:334] "Generic (PLEG): container finished" podID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerID="2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4" exitCode=0 Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.146246 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerDied","Data":"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4"} Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.146269 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q5t7p" event={"ID":"48f780d0-874f-4d5e-b48e-79abb6cf0e28","Type":"ContainerDied","Data":"4efa6b5685bd4815fc5baf0bc2f59a384a75e9667820bb77e614363860246fd5"} Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.146326 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q5t7p" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184479 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqnjz\" (UniqueName: \"kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz\") pod \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184602 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities\") pod \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184667 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content\") pod \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184729 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities\") pod \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184772 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content\") pod \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\" (UID: \"2e6078c0-b5c1-488d-9ed3-ff31fb87dd49\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.184788 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmg8z\" (UniqueName: \"kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z\") pod \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\" (UID: \"48f780d0-874f-4d5e-b48e-79abb6cf0e28\") " Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.185601 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities" (OuterVolumeSpecName: "utilities") pod "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" (UID: "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.185672 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities" (OuterVolumeSpecName: "utilities") pod "48f780d0-874f-4d5e-b48e-79abb6cf0e28" (UID: "48f780d0-874f-4d5e-b48e-79abb6cf0e28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.186655 4983 scope.go:117] "RemoveContainer" containerID="58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.190042 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz" (OuterVolumeSpecName: "kube-api-access-jqnjz") pod "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" (UID: "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49"). InnerVolumeSpecName "kube-api-access-jqnjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.190331 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z" (OuterVolumeSpecName: "kube-api-access-nmg8z") pod "48f780d0-874f-4d5e-b48e-79abb6cf0e28" (UID: "48f780d0-874f-4d5e-b48e-79abb6cf0e28"). InnerVolumeSpecName "kube-api-access-nmg8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.227351 4983 scope.go:117] "RemoveContainer" containerID="2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.248614 4983 scope.go:117] "RemoveContainer" containerID="d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.249089 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6\": container with ID starting with d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6 not found: ID does not exist" containerID="d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.249118 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6"} err="failed to get container status \"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6\": rpc error: code = NotFound desc = could not find container \"d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6\": container with ID starting with d1677bb8eeca8a494df702bf49de75e9c5f3491b0e4cc5b11225496295a15db6 not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.249139 4983 scope.go:117] "RemoveContainer" containerID="58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.250079 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716\": container with ID starting with 58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716 not found: ID does not exist" containerID="58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.250120 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716"} err="failed to get container status \"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716\": rpc error: code = NotFound desc = could not find container \"58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716\": container with ID starting with 58f72e768a0a9132cfba85a7b33085fb350ffcc2c196808015cdd0e217c48716 not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.250147 4983 scope.go:117] "RemoveContainer" containerID="2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.250492 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb\": container with ID starting with 2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb not found: ID does not exist" containerID="2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.250521 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb"} err="failed to get container status \"2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb\": rpc error: code = NotFound desc = could not find container \"2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb\": container with ID starting with 2150fc20c59910f6970fc92b392a4cabc21ec503837f08cbf7d250209774b5fb not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.250536 4983 scope.go:117] "RemoveContainer" containerID="2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.252861 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" (UID: "2e6078c0-b5c1-488d-9ed3-ff31fb87dd49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.254551 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48f780d0-874f-4d5e-b48e-79abb6cf0e28" (UID: "48f780d0-874f-4d5e-b48e-79abb6cf0e28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.263664 4983 scope.go:117] "RemoveContainer" containerID="ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286157 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqnjz\" (UniqueName: \"kubernetes.io/projected/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-kube-api-access-jqnjz\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286189 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286204 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48f780d0-874f-4d5e-b48e-79abb6cf0e28-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286215 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286226 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.286299 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmg8z\" (UniqueName: \"kubernetes.io/projected/48f780d0-874f-4d5e-b48e-79abb6cf0e28-kube-api-access-nmg8z\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.305439 4983 scope.go:117] "RemoveContainer" containerID="a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.320093 4983 scope.go:117] "RemoveContainer" containerID="2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.320789 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4\": container with ID starting with 2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4 not found: ID does not exist" containerID="2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.320965 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4"} err="failed to get container status \"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4\": rpc error: code = NotFound desc = could not find container \"2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4\": container with ID starting with 2852b0cee5e4801b6b11d386d28d6a722392535bd5f4ad950343590ec816ded4 not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.321052 4983 scope.go:117] "RemoveContainer" containerID="ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.321518 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a\": container with ID starting with ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a not found: ID does not exist" containerID="ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.321557 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a"} err="failed to get container status \"ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a\": rpc error: code = NotFound desc = could not find container \"ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a\": container with ID starting with ca0dad234afa4bb3a64760888aa764a0da1f9bf414c5da621cef8a5f3424002a not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.321585 4983 scope.go:117] "RemoveContainer" containerID="a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56" Oct 01 09:13:02 crc kubenswrapper[4983]: E1001 09:13:02.321903 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56\": container with ID starting with a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56 not found: ID does not exist" containerID="a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.321960 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56"} err="failed to get container status \"a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56\": rpc error: code = NotFound desc = could not find container \"a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56\": container with ID starting with a7c6fb0a120c2741321cc5269fc0e88cd72d8a140dcbd4f47f14a3e253134f56 not found: ID does not exist" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.466905 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.470248 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-22tvh"] Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.476520 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.480207 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q5t7p"] Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.720492 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" path="/var/lib/kubelet/pods/2e6078c0-b5c1-488d-9ed3-ff31fb87dd49/volumes" Oct 01 09:13:02 crc kubenswrapper[4983]: I1001 09:13:02.721455 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" path="/var/lib/kubelet/pods/48f780d0-874f-4d5e-b48e-79abb6cf0e28/volumes" Oct 01 09:13:03 crc kubenswrapper[4983]: I1001 09:13:03.152842 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3"} Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.162145 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.162350 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l5882" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="registry-server" containerID="cri-o://6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00" gracePeriod=2 Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.502200 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.614826 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content\") pod \"2f86c554-fecb-48e6-b257-8f398bdbc97d\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.614891 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities\") pod \"2f86c554-fecb-48e6-b257-8f398bdbc97d\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.614938 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz5hp\" (UniqueName: \"kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp\") pod \"2f86c554-fecb-48e6-b257-8f398bdbc97d\" (UID: \"2f86c554-fecb-48e6-b257-8f398bdbc97d\") " Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.615928 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities" (OuterVolumeSpecName: "utilities") pod "2f86c554-fecb-48e6-b257-8f398bdbc97d" (UID: "2f86c554-fecb-48e6-b257-8f398bdbc97d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.620661 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp" (OuterVolumeSpecName: "kube-api-access-nz5hp") pod "2f86c554-fecb-48e6-b257-8f398bdbc97d" (UID: "2f86c554-fecb-48e6-b257-8f398bdbc97d"). InnerVolumeSpecName "kube-api-access-nz5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.629527 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f86c554-fecb-48e6-b257-8f398bdbc97d" (UID: "2f86c554-fecb-48e6-b257-8f398bdbc97d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.716511 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.716547 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f86c554-fecb-48e6-b257-8f398bdbc97d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:04 crc kubenswrapper[4983]: I1001 09:13:04.716560 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz5hp\" (UniqueName: \"kubernetes.io/projected/2f86c554-fecb-48e6-b257-8f398bdbc97d-kube-api-access-nz5hp\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.167260 4983 generic.go:334] "Generic (PLEG): container finished" podID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerID="6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00" exitCode=0 Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.167313 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerDied","Data":"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00"} Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.167292 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5882" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.167348 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5882" event={"ID":"2f86c554-fecb-48e6-b257-8f398bdbc97d","Type":"ContainerDied","Data":"7a39c423a140e6e3124b0ded323f50cca882a8532fb6fc6e55789a22b2fe299b"} Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.167366 4983 scope.go:117] "RemoveContainer" containerID="6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.182162 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.186130 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5882"] Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.189299 4983 scope.go:117] "RemoveContainer" containerID="02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.203984 4983 scope.go:117] "RemoveContainer" containerID="a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.229750 4983 scope.go:117] "RemoveContainer" containerID="6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00" Oct 01 09:13:05 crc kubenswrapper[4983]: E1001 09:13:05.230313 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00\": container with ID starting with 6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00 not found: ID does not exist" containerID="6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.230390 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00"} err="failed to get container status \"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00\": rpc error: code = NotFound desc = could not find container \"6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00\": container with ID starting with 6d2b768e00695c93b372545801a154b3703945a5f1c338bf52e5b945fb1ffd00 not found: ID does not exist" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.230440 4983 scope.go:117] "RemoveContainer" containerID="02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185" Oct 01 09:13:05 crc kubenswrapper[4983]: E1001 09:13:05.231160 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185\": container with ID starting with 02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185 not found: ID does not exist" containerID="02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.231190 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185"} err="failed to get container status \"02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185\": rpc error: code = NotFound desc = could not find container \"02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185\": container with ID starting with 02bdf050e9f8f40699a6a8d4a715a1d7b1f604780cc8fed590be1b28a2132185 not found: ID does not exist" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.231211 4983 scope.go:117] "RemoveContainer" containerID="a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6" Oct 01 09:13:05 crc kubenswrapper[4983]: E1001 09:13:05.231616 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6\": container with ID starting with a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6 not found: ID does not exist" containerID="a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6" Oct 01 09:13:05 crc kubenswrapper[4983]: I1001 09:13:05.231738 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6"} err="failed to get container status \"a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6\": rpc error: code = NotFound desc = could not find container \"a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6\": container with ID starting with a0e3dbf43249f01b1161dffdd6caa42e469d28fb6e6fd06e7ea0e5e947654ad6 not found: ID does not exist" Oct 01 09:13:06 crc kubenswrapper[4983]: I1001 09:13:06.719918 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" path="/var/lib/kubelet/pods/2f86c554-fecb-48e6-b257-8f398bdbc97d/volumes" Oct 01 09:13:12 crc kubenswrapper[4983]: I1001 09:13:12.261095 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.301295 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerName="oauth-openshift" containerID="cri-o://9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c" gracePeriod=15 Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.780211 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820097 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-z87kb"] Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820440 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820475 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820506 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820522 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820543 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75061baf-2309-4dcc-be5b-eb39bff7dfea" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820557 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="75061baf-2309-4dcc-be5b-eb39bff7dfea" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820577 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820590 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820610 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53fd5c81-7bb2-4411-8d0c-9a0ad13da649" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820623 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="53fd5c81-7bb2-4411-8d0c-9a0ad13da649" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820643 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820656 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820671 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerName="oauth-openshift" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820683 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerName="oauth-openshift" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820703 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820716 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820736 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820750 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820768 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820780 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820802 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820855 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820880 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820893 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820911 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820925 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820948 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820960 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="extract-utilities" Oct 01 09:13:37 crc kubenswrapper[4983]: E1001 09:13:37.820975 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.820987 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="extract-content" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821182 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="53fd5c81-7bb2-4411-8d0c-9a0ad13da649" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821211 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerName="oauth-openshift" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821227 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6078c0-b5c1-488d-9ed3-ff31fb87dd49" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821246 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f780d0-874f-4d5e-b48e-79abb6cf0e28" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821267 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf90392d-40b6-400b-941b-0fcd6bc8b30b" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821287 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="75061baf-2309-4dcc-be5b-eb39bff7dfea" containerName="pruner" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821303 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f86c554-fecb-48e6-b257-8f398bdbc97d" containerName="registry-server" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.821989 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.837856 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-z87kb"] Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838187 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbh98\" (UniqueName: \"kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838239 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838272 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838339 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838384 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838412 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838439 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838460 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838477 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838508 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838548 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838567 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838587 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data\") pod \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\" (UID: \"e5f07bb1-dbb5-4a9c-90ed-caec428cd247\") " Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838719 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838749 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24m4s\" (UniqueName: \"kubernetes.io/projected/8083bbb5-5779-452c-a831-6c75a43204ea-kube-api-access-24m4s\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838769 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838790 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838829 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838851 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838892 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8083bbb5-5779-452c-a831-6c75a43204ea-audit-dir\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838934 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838966 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.838993 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.839009 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.839037 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.839067 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-audit-policies\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.839088 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.840677 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.841583 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.841772 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.847416 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.847127 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.848839 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.849997 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98" (OuterVolumeSpecName: "kube-api-access-fbh98") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "kube-api-access-fbh98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.851307 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.852350 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.857197 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.860724 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.864444 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.864753 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.865086 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e5f07bb1-dbb5-4a9c-90ed-caec428cd247" (UID: "e5f07bb1-dbb5-4a9c-90ed-caec428cd247"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.940401 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.940703 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.940851 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.940936 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941015 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941092 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-audit-policies\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941171 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941239 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941318 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24m4s\" (UniqueName: \"kubernetes.io/projected/8083bbb5-5779-452c-a831-6c75a43204ea-kube-api-access-24m4s\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941393 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941492 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941567 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941636 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941717 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8083bbb5-5779-452c-a831-6c75a43204ea-audit-dir\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.941845 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbh98\" (UniqueName: \"kubernetes.io/projected/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-kube-api-access-fbh98\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.942095 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.942740 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-audit-policies\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.942945 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8083bbb5-5779-452c-a831-6c75a43204ea-audit-dir\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943166 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943069 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943347 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943360 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943370 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943383 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943393 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943393 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943404 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943480 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943499 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943519 4983 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943535 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943550 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e5f07bb1-dbb5-4a9c-90ed-caec428cd247-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.943644 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.944707 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.944886 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.945780 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.945782 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.946600 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.946844 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.947065 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.947913 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8083bbb5-5779-452c-a831-6c75a43204ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:37 crc kubenswrapper[4983]: I1001 09:13:37.958125 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24m4s\" (UniqueName: \"kubernetes.io/projected/8083bbb5-5779-452c-a831-6c75a43204ea-kube-api-access-24m4s\") pod \"oauth-openshift-57569d6b9d-z87kb\" (UID: \"8083bbb5-5779-452c-a831-6c75a43204ea\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.188089 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.375442 4983 generic.go:334] "Generic (PLEG): container finished" podID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" containerID="9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c" exitCode=0 Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.375532 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.375514 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" event={"ID":"e5f07bb1-dbb5-4a9c-90ed-caec428cd247","Type":"ContainerDied","Data":"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c"} Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.375915 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z72rd" event={"ID":"e5f07bb1-dbb5-4a9c-90ed-caec428cd247","Type":"ContainerDied","Data":"44c91d8db86a9c701f7b0ae9a8d3baa00639d6f7d65a8047bb9e35632dc10e74"} Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.375943 4983 scope.go:117] "RemoveContainer" containerID="9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.413205 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.414507 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z72rd"] Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.430667 4983 scope.go:117] "RemoveContainer" containerID="9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c" Oct 01 09:13:38 crc kubenswrapper[4983]: E1001 09:13:38.431603 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c\": container with ID starting with 9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c not found: ID does not exist" containerID="9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.431632 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c"} err="failed to get container status \"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c\": rpc error: code = NotFound desc = could not find container \"9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c\": container with ID starting with 9ed1b640bb0fbc20160f39bcfcfdf7d8c1de1b4b426b18ffaeee1e36db29d92c not found: ID does not exist" Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.441084 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-z87kb"] Oct 01 09:13:38 crc kubenswrapper[4983]: W1001 09:13:38.451238 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8083bbb5_5779_452c_a831_6c75a43204ea.slice/crio-1169289a5726f8190ef2fe9837547d5a9660a2d1f6b66c3d7a47471758bb7f61 WatchSource:0}: Error finding container 1169289a5726f8190ef2fe9837547d5a9660a2d1f6b66c3d7a47471758bb7f61: Status 404 returned error can't find the container with id 1169289a5726f8190ef2fe9837547d5a9660a2d1f6b66c3d7a47471758bb7f61 Oct 01 09:13:38 crc kubenswrapper[4983]: I1001 09:13:38.726294 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f07bb1-dbb5-4a9c-90ed-caec428cd247" path="/var/lib/kubelet/pods/e5f07bb1-dbb5-4a9c-90ed-caec428cd247/volumes" Oct 01 09:13:39 crc kubenswrapper[4983]: I1001 09:13:39.386748 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" event={"ID":"8083bbb5-5779-452c-a831-6c75a43204ea","Type":"ContainerStarted","Data":"9b72f72f03e0e203629a81d1c5983b575c042abc639401b845e5fe92d3000e75"} Oct 01 09:13:39 crc kubenswrapper[4983]: I1001 09:13:39.387176 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" event={"ID":"8083bbb5-5779-452c-a831-6c75a43204ea","Type":"ContainerStarted","Data":"1169289a5726f8190ef2fe9837547d5a9660a2d1f6b66c3d7a47471758bb7f61"} Oct 01 09:13:39 crc kubenswrapper[4983]: I1001 09:13:39.388704 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:39 crc kubenswrapper[4983]: I1001 09:13:39.402648 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" Oct 01 09:13:39 crc kubenswrapper[4983]: I1001 09:13:39.420558 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57569d6b9d-z87kb" podStartSLOduration=27.420532025 podStartE2EDuration="27.420532025s" podCreationTimestamp="2025-10-01 09:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:13:39.417521725 +0000 UTC m=+247.406750552" watchObservedRunningTime="2025-10-01 09:13:39.420532025 +0000 UTC m=+247.409760832" Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.871522 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.872432 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2lvtf" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="registry-server" containerID="cri-o://1f8637028c7e6da5f8e78214336d4d3a9fd1a74bc728665818df0d1b8cbfc53c" gracePeriod=30 Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.888906 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.891968 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.892032 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-clr5p" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="registry-server" containerID="cri-o://e9565826580afc48bc82ead6595de1dd0da52ab850e186bd35db54668b827706" gracePeriod=30 Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.892193 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" containerID="cri-o://3585414e687f49642614de54551b6dbfc57f3c318f4752fd0415498543390460" gracePeriod=30 Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.913141 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.913755 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ptxwd" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="registry-server" containerID="cri-o://1cf101d20ea9972ece00fe7940d38c67aa046aaa835a77fdf3f2434ec8931154" gracePeriod=30 Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.915237 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddvpk"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.915958 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.917390 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.917626 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cdcrl" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" containerID="cri-o://1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" gracePeriod=30 Oct 01 09:13:51 crc kubenswrapper[4983]: I1001 09:13:51.925486 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddvpk"] Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.027602 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.027698 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.027725 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lplft\" (UniqueName: \"kubernetes.io/projected/5d2dbdcc-216c-4e21-8da0-3291670f7555-kube-api-access-lplft\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.129149 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.129225 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.129248 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lplft\" (UniqueName: \"kubernetes.io/projected/5d2dbdcc-216c-4e21-8da0-3291670f7555-kube-api-access-lplft\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.130382 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.137409 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d2dbdcc-216c-4e21-8da0-3291670f7555-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.144457 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lplft\" (UniqueName: \"kubernetes.io/projected/5d2dbdcc-216c-4e21-8da0-3291670f7555-kube-api-access-lplft\") pod \"marketplace-operator-79b997595-ddvpk\" (UID: \"5d2dbdcc-216c-4e21-8da0-3291670f7555\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.234205 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:52 crc kubenswrapper[4983]: E1001 09:13:52.348924 4983 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1 is running failed: container process not found" containerID="1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 09:13:52 crc kubenswrapper[4983]: E1001 09:13:52.349439 4983 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1 is running failed: container process not found" containerID="1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 09:13:52 crc kubenswrapper[4983]: E1001 09:13:52.349817 4983 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1 is running failed: container process not found" containerID="1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 09:13:52 crc kubenswrapper[4983]: E1001 09:13:52.349852 4983 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-cdcrl" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.466935 4983 generic.go:334] "Generic (PLEG): container finished" podID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerID="e9565826580afc48bc82ead6595de1dd0da52ab850e186bd35db54668b827706" exitCode=0 Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.467014 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerDied","Data":"e9565826580afc48bc82ead6595de1dd0da52ab850e186bd35db54668b827706"} Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.469246 4983 generic.go:334] "Generic (PLEG): container finished" podID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerID="1f8637028c7e6da5f8e78214336d4d3a9fd1a74bc728665818df0d1b8cbfc53c" exitCode=0 Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.469299 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerDied","Data":"1f8637028c7e6da5f8e78214336d4d3a9fd1a74bc728665818df0d1b8cbfc53c"} Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.472346 4983 generic.go:334] "Generic (PLEG): container finished" podID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerID="1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" exitCode=0 Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.472412 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerDied","Data":"1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1"} Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.473958 4983 generic.go:334] "Generic (PLEG): container finished" podID="a199e85c-98a8-4563-81a6-196998a19005" containerID="3585414e687f49642614de54551b6dbfc57f3c318f4752fd0415498543390460" exitCode=0 Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.473992 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" event={"ID":"a199e85c-98a8-4563-81a6-196998a19005","Type":"ContainerDied","Data":"3585414e687f49642614de54551b6dbfc57f3c318f4752fd0415498543390460"} Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.475857 4983 generic.go:334] "Generic (PLEG): container finished" podID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerID="1cf101d20ea9972ece00fe7940d38c67aa046aaa835a77fdf3f2434ec8931154" exitCode=0 Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.475877 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerDied","Data":"1cf101d20ea9972ece00fe7940d38c67aa046aaa835a77fdf3f2434ec8931154"} Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.601707 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddvpk"] Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.731361 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.804127 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843135 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") pod \"a199e85c-98a8-4563-81a6-196998a19005\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843185 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities\") pod \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843226 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnw9x\" (UniqueName: \"kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x\") pod \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843267 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9v7f\" (UniqueName: \"kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f\") pod \"a199e85c-98a8-4563-81a6-196998a19005\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843305 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") pod \"a199e85c-98a8-4563-81a6-196998a19005\" (UID: \"a199e85c-98a8-4563-81a6-196998a19005\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.843356 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content\") pod \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\" (UID: \"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.845099 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a199e85c-98a8-4563-81a6-196998a19005" (UID: "a199e85c-98a8-4563-81a6-196998a19005"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.845791 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities" (OuterVolumeSpecName: "utilities") pod "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" (UID: "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.849885 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x" (OuterVolumeSpecName: "kube-api-access-nnw9x") pod "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" (UID: "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79"). InnerVolumeSpecName "kube-api-access-nnw9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.850181 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f" (OuterVolumeSpecName: "kube-api-access-p9v7f") pod "a199e85c-98a8-4563-81a6-196998a19005" (UID: "a199e85c-98a8-4563-81a6-196998a19005"). InnerVolumeSpecName "kube-api-access-p9v7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.851266 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a199e85c-98a8-4563-81a6-196998a19005" (UID: "a199e85c-98a8-4563-81a6-196998a19005"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.866794 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.875602 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.899569 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944102 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content\") pod \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944147 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content\") pod \"b9f5a446-fb85-404e-9200-f38d58f5442f\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities\") pod \"b9f5a446-fb85-404e-9200-f38d58f5442f\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944227 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities\") pod \"be0c00c3-40f3-4806-bb61-af2dee4dd318\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944242 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities\") pod \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944296 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content\") pod \"be0c00c3-40f3-4806-bb61-af2dee4dd318\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944317 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lttbm\" (UniqueName: \"kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm\") pod \"b9f5a446-fb85-404e-9200-f38d58f5442f\" (UID: \"b9f5a446-fb85-404e-9200-f38d58f5442f\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944341 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qszfz\" (UniqueName: \"kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz\") pod \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\" (UID: \"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944357 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bprlf\" (UniqueName: \"kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf\") pod \"be0c00c3-40f3-4806-bb61-af2dee4dd318\" (UID: \"be0c00c3-40f3-4806-bb61-af2dee4dd318\") " Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944553 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a199e85c-98a8-4563-81a6-196998a19005-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944567 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a199e85c-98a8-4563-81a6-196998a19005-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944576 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944586 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnw9x\" (UniqueName: \"kubernetes.io/projected/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-kube-api-access-nnw9x\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.944596 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9v7f\" (UniqueName: \"kubernetes.io/projected/a199e85c-98a8-4563-81a6-196998a19005-kube-api-access-p9v7f\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.948329 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities" (OuterVolumeSpecName: "utilities") pod "b9f5a446-fb85-404e-9200-f38d58f5442f" (UID: "b9f5a446-fb85-404e-9200-f38d58f5442f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.949111 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities" (OuterVolumeSpecName: "utilities") pod "be0c00c3-40f3-4806-bb61-af2dee4dd318" (UID: "be0c00c3-40f3-4806-bb61-af2dee4dd318"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.951964 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm" (OuterVolumeSpecName: "kube-api-access-lttbm") pod "b9f5a446-fb85-404e-9200-f38d58f5442f" (UID: "b9f5a446-fb85-404e-9200-f38d58f5442f"). InnerVolumeSpecName "kube-api-access-lttbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.954920 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities" (OuterVolumeSpecName: "utilities") pod "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" (UID: "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.956974 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz" (OuterVolumeSpecName: "kube-api-access-qszfz") pod "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" (UID: "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3"). InnerVolumeSpecName "kube-api-access-qszfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.962202 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf" (OuterVolumeSpecName: "kube-api-access-bprlf") pod "be0c00c3-40f3-4806-bb61-af2dee4dd318" (UID: "be0c00c3-40f3-4806-bb61-af2dee4dd318"). InnerVolumeSpecName "kube-api-access-bprlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.980700 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" (UID: "8a77c4f0-1ce1-45b4-8e9a-3c7360290e79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:52 crc kubenswrapper[4983]: I1001 09:13:52.990255 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" (UID: "ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.037547 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9f5a446-fb85-404e-9200-f38d58f5442f" (UID: "b9f5a446-fb85-404e-9200-f38d58f5442f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045497 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045525 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045533 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045542 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045550 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lttbm\" (UniqueName: \"kubernetes.io/projected/b9f5a446-fb85-404e-9200-f38d58f5442f-kube-api-access-lttbm\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045559 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qszfz\" (UniqueName: \"kubernetes.io/projected/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-kube-api-access-qszfz\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045567 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bprlf\" (UniqueName: \"kubernetes.io/projected/be0c00c3-40f3-4806-bb61-af2dee4dd318-kube-api-access-bprlf\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045575 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f5a446-fb85-404e-9200-f38d58f5442f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.045583 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.060515 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be0c00c3-40f3-4806-bb61-af2dee4dd318" (UID: "be0c00c3-40f3-4806-bb61-af2dee4dd318"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.147294 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be0c00c3-40f3-4806-bb61-af2dee4dd318-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.482050 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lvtf" event={"ID":"8a77c4f0-1ce1-45b4-8e9a-3c7360290e79","Type":"ContainerDied","Data":"bcd6f7b86d1f2b53eb2639f04a608422a3a626aa86fbe2cfe19b83bf0baed6f2"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.482098 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lvtf" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.482117 4983 scope.go:117] "RemoveContainer" containerID="1f8637028c7e6da5f8e78214336d4d3a9fd1a74bc728665818df0d1b8cbfc53c" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.486040 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cdcrl" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.486025 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cdcrl" event={"ID":"be0c00c3-40f3-4806-bb61-af2dee4dd318","Type":"ContainerDied","Data":"bc75a305119b6ae84a2e3bd95bd924f5956d3bfa0a91e3af9cc29f220f4cd058"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.488290 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" event={"ID":"a199e85c-98a8-4563-81a6-196998a19005","Type":"ContainerDied","Data":"3ed738223af0fc9a61d526e6bd2f8f0feaf2456be8a3e3e9d07d9f157cb56132"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.488303 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xhsfh" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.491217 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ptxwd" event={"ID":"ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3","Type":"ContainerDied","Data":"0f09f1295ecf84fc7944ca88b0f77429eccc720724af2d80a4e04bedf5b50bcc"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.491356 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ptxwd" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.496064 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" event={"ID":"5d2dbdcc-216c-4e21-8da0-3291670f7555","Type":"ContainerStarted","Data":"e03d484001e98bdea6c3d9af50c001cc330ba85a22160416c74c88021e82d807"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.496121 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" event={"ID":"5d2dbdcc-216c-4e21-8da0-3291670f7555","Type":"ContainerStarted","Data":"e1b4c047c2110ba46bea33bd3ae052e9d4a3b005b2bec624ee1b34606a061238"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.496270 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.498280 4983 scope.go:117] "RemoveContainer" containerID="bafae28a91eb2faa1e632e679bc654ca53542500ac41c5b863b862df6c296fae" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.501087 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.501175 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clr5p" event={"ID":"b9f5a446-fb85-404e-9200-f38d58f5442f","Type":"ContainerDied","Data":"8aee3124f860a5c5e3c3761350aac8c7aa0ea883ed0cbc51a38728069bec1684"} Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.501250 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clr5p" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.521740 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ddvpk" podStartSLOduration=2.521714718 podStartE2EDuration="2.521714718s" podCreationTimestamp="2025-10-01 09:13:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:13:53.515871603 +0000 UTC m=+261.505100420" watchObservedRunningTime="2025-10-01 09:13:53.521714718 +0000 UTC m=+261.510943515" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.529232 4983 scope.go:117] "RemoveContainer" containerID="c1c9c82291b10d4386f6367c6920f84424e6e8663b2f3ff54853f8aed9cd36ac" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.535240 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.539250 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cdcrl"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.546470 4983 scope.go:117] "RemoveContainer" containerID="1a9576fc65e70b97bb36e52b999390fddf84f9c26783ef48de1eeea86d971fb1" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.566770 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.569159 4983 scope.go:117] "RemoveContainer" containerID="ee659ae610366ae39530b905c08421420d6c2a34ef7a73d3e2cf29cf892c58fd" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.577914 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2lvtf"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.578675 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.587110 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ptxwd"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.588448 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.590870 4983 scope.go:117] "RemoveContainer" containerID="354db233e23a8c6c5a9c780ab22d3b7c037eb4f91dc1de438865ebbbbf39b00a" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.598434 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-clr5p"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.601871 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.603528 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xhsfh"] Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.618286 4983 scope.go:117] "RemoveContainer" containerID="3585414e687f49642614de54551b6dbfc57f3c318f4752fd0415498543390460" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.643294 4983 scope.go:117] "RemoveContainer" containerID="1cf101d20ea9972ece00fe7940d38c67aa046aaa835a77fdf3f2434ec8931154" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.656144 4983 scope.go:117] "RemoveContainer" containerID="2cd655e71335624596cb1aa2dd7b70ca4fd9eb5a5623394ee30232b765c834ed" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.677626 4983 scope.go:117] "RemoveContainer" containerID="413d4fd2ca9d96721f056e97de7901459899b3e6198468266d6c430dd771a673" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.696781 4983 scope.go:117] "RemoveContainer" containerID="e9565826580afc48bc82ead6595de1dd0da52ab850e186bd35db54668b827706" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.712449 4983 scope.go:117] "RemoveContainer" containerID="0287ffe159d3e7502e7b58f4c77872edb4d82d1103a05cf76944e16f91a6f8ba" Oct 01 09:13:53 crc kubenswrapper[4983]: I1001 09:13:53.732107 4983 scope.go:117] "RemoveContainer" containerID="49e3b784eb98dc25dceb501f9297174669112e978833d4271c62efe3b90bc4c3" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099354 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8lq6m"] Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099640 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099653 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099663 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099670 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099678 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099684 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099691 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099717 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099724 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099730 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099739 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099745 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099754 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099759 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099766 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099771 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099830 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099838 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099846 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099852 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099861 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099866 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099873 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099878 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="extract-utilities" Oct 01 09:13:54 crc kubenswrapper[4983]: E1001 09:13:54.099886 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.099932 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="extract-content" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.100042 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.102559 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a199e85c-98a8-4563-81a6-196998a19005" containerName="marketplace-operator" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.102576 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.102585 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.102596 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" containerName="registry-server" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.103174 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lq6m"] Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.103482 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.105522 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.156131 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-utilities\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.156193 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrrh\" (UniqueName: \"kubernetes.io/projected/62d5d2de-567c-4077-8d82-32833bfb6129-kube-api-access-nxrrh\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.156371 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-catalog-content\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.257281 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-utilities\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.257333 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrrh\" (UniqueName: \"kubernetes.io/projected/62d5d2de-567c-4077-8d82-32833bfb6129-kube-api-access-nxrrh\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.257376 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-catalog-content\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.257925 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-utilities\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.257952 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62d5d2de-567c-4077-8d82-32833bfb6129-catalog-content\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.278925 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrrh\" (UniqueName: \"kubernetes.io/projected/62d5d2de-567c-4077-8d82-32833bfb6129-kube-api-access-nxrrh\") pod \"redhat-marketplace-8lq6m\" (UID: \"62d5d2de-567c-4077-8d82-32833bfb6129\") " pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.300675 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fcr4m"] Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.302264 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.304169 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.306736 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fcr4m"] Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.358098 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-utilities\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.358143 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgttn\" (UniqueName: \"kubernetes.io/projected/0ccb3b8a-516c-4c88-8350-21a238b80504-kube-api-access-bgttn\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.358168 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-catalog-content\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.421104 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.459001 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-catalog-content\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.459100 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-utilities\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.459125 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgttn\" (UniqueName: \"kubernetes.io/projected/0ccb3b8a-516c-4c88-8350-21a238b80504-kube-api-access-bgttn\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.459836 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-catalog-content\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.460066 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ccb3b8a-516c-4c88-8350-21a238b80504-utilities\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.474676 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgttn\" (UniqueName: \"kubernetes.io/projected/0ccb3b8a-516c-4c88-8350-21a238b80504-kube-api-access-bgttn\") pod \"redhat-operators-fcr4m\" (UID: \"0ccb3b8a-516c-4c88-8350-21a238b80504\") " pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.582351 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lq6m"] Oct 01 09:13:54 crc kubenswrapper[4983]: W1001 09:13:54.587945 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62d5d2de_567c_4077_8d82_32833bfb6129.slice/crio-8c57768e958bd84dfc3723329e804c5dca37b8b52b90a07a6c3462b9bd8ed1ce WatchSource:0}: Error finding container 8c57768e958bd84dfc3723329e804c5dca37b8b52b90a07a6c3462b9bd8ed1ce: Status 404 returned error can't find the container with id 8c57768e958bd84dfc3723329e804c5dca37b8b52b90a07a6c3462b9bd8ed1ce Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.624240 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.721473 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a77c4f0-1ce1-45b4-8e9a-3c7360290e79" path="/var/lib/kubelet/pods/8a77c4f0-1ce1-45b4-8e9a-3c7360290e79/volumes" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.722267 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a199e85c-98a8-4563-81a6-196998a19005" path="/var/lib/kubelet/pods/a199e85c-98a8-4563-81a6-196998a19005/volumes" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.722801 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3" path="/var/lib/kubelet/pods/ae86b8ef-7a6d-4cb7-a9f0-3bdad92088a3/volumes" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.724293 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f5a446-fb85-404e-9200-f38d58f5442f" path="/var/lib/kubelet/pods/b9f5a446-fb85-404e-9200-f38d58f5442f/volumes" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.725281 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be0c00c3-40f3-4806-bb61-af2dee4dd318" path="/var/lib/kubelet/pods/be0c00c3-40f3-4806-bb61-af2dee4dd318/volumes" Oct 01 09:13:54 crc kubenswrapper[4983]: I1001 09:13:54.807028 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fcr4m"] Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.521781 4983 generic.go:334] "Generic (PLEG): container finished" podID="62d5d2de-567c-4077-8d82-32833bfb6129" containerID="844769bd00d67cfc988a442c73f153ab5bb54f95f64ced226abcf8df36c71505" exitCode=0 Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.521847 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lq6m" event={"ID":"62d5d2de-567c-4077-8d82-32833bfb6129","Type":"ContainerDied","Data":"844769bd00d67cfc988a442c73f153ab5bb54f95f64ced226abcf8df36c71505"} Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.522191 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lq6m" event={"ID":"62d5d2de-567c-4077-8d82-32833bfb6129","Type":"ContainerStarted","Data":"8c57768e958bd84dfc3723329e804c5dca37b8b52b90a07a6c3462b9bd8ed1ce"} Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.524319 4983 generic.go:334] "Generic (PLEG): container finished" podID="0ccb3b8a-516c-4c88-8350-21a238b80504" containerID="a4e2bdbb895469714ef568fc86ec97ebc6cf5cae48fa5292cc3e1e72f40c74bf" exitCode=0 Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.524388 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fcr4m" event={"ID":"0ccb3b8a-516c-4c88-8350-21a238b80504","Type":"ContainerDied","Data":"a4e2bdbb895469714ef568fc86ec97ebc6cf5cae48fa5292cc3e1e72f40c74bf"} Oct 01 09:13:55 crc kubenswrapper[4983]: I1001 09:13:55.524428 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fcr4m" event={"ID":"0ccb3b8a-516c-4c88-8350-21a238b80504","Type":"ContainerStarted","Data":"ff5ad71de15437e8005a858fc33ceeb95f5ebcb458f47529ad704bf95d3b885b"} Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.486034 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4f8xr"] Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.488712 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.490380 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.497372 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4f8xr"] Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.529508 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fcr4m" event={"ID":"0ccb3b8a-516c-4c88-8350-21a238b80504","Type":"ContainerStarted","Data":"1a58e4ecad95a83c38db98cf957bb9796479c97896fc07969ab0bab68214214f"} Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.532336 4983 generic.go:334] "Generic (PLEG): container finished" podID="62d5d2de-567c-4077-8d82-32833bfb6129" containerID="6565046f4e6a2373235a8ebd277d8486d14c98fa69662eb126c3a4cc8044bc7e" exitCode=0 Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.532370 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lq6m" event={"ID":"62d5d2de-567c-4077-8d82-32833bfb6129","Type":"ContainerDied","Data":"6565046f4e6a2373235a8ebd277d8486d14c98fa69662eb126c3a4cc8044bc7e"} Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.581550 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-catalog-content\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.581687 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-utilities\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.581755 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2stj\" (UniqueName: \"kubernetes.io/projected/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-kube-api-access-q2stj\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.682555 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-utilities\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.682965 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2stj\" (UniqueName: \"kubernetes.io/projected/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-kube-api-access-q2stj\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.683027 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-catalog-content\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.684111 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-catalog-content\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.684253 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-utilities\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.689612 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cg4hg"] Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.691239 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.704179 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cg4hg"] Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.704936 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.713982 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2stj\" (UniqueName: \"kubernetes.io/projected/b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63-kube-api-access-q2stj\") pod \"certified-operators-4f8xr\" (UID: \"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63\") " pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.783847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-utilities\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.783900 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-catalog-content\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.783953 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j76r\" (UniqueName: \"kubernetes.io/projected/c982d336-6845-4efb-8a7b-202f55f55e8a-kube-api-access-7j76r\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.843094 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.884872 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-utilities\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.884923 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-catalog-content\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.884959 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j76r\" (UniqueName: \"kubernetes.io/projected/c982d336-6845-4efb-8a7b-202f55f55e8a-kube-api-access-7j76r\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.885395 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-utilities\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.885520 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c982d336-6845-4efb-8a7b-202f55f55e8a-catalog-content\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:56 crc kubenswrapper[4983]: I1001 09:13:56.902440 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j76r\" (UniqueName: \"kubernetes.io/projected/c982d336-6845-4efb-8a7b-202f55f55e8a-kube-api-access-7j76r\") pod \"community-operators-cg4hg\" (UID: \"c982d336-6845-4efb-8a7b-202f55f55e8a\") " pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.006999 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.025291 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4f8xr"] Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.211545 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cg4hg"] Oct 01 09:13:57 crc kubenswrapper[4983]: W1001 09:13:57.230496 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc982d336_6845_4efb_8a7b_202f55f55e8a.slice/crio-3aec9a2a946dd7c2fe00b5994e4cc385a346c54b5d7a10e1b691eb207868c306 WatchSource:0}: Error finding container 3aec9a2a946dd7c2fe00b5994e4cc385a346c54b5d7a10e1b691eb207868c306: Status 404 returned error can't find the container with id 3aec9a2a946dd7c2fe00b5994e4cc385a346c54b5d7a10e1b691eb207868c306 Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.541667 4983 generic.go:334] "Generic (PLEG): container finished" podID="b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63" containerID="d497ec83378ee5b3882baf85b6dbdf08c1c033ad620e0c08dc8238330f0d6a7e" exitCode=0 Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.541742 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4f8xr" event={"ID":"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63","Type":"ContainerDied","Data":"d497ec83378ee5b3882baf85b6dbdf08c1c033ad620e0c08dc8238330f0d6a7e"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.541770 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4f8xr" event={"ID":"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63","Type":"ContainerStarted","Data":"58d34789e31fa23bf243cb28ab66b6fee2a5c5e86430e5a6f78be4bba4dba678"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.543736 4983 generic.go:334] "Generic (PLEG): container finished" podID="c982d336-6845-4efb-8a7b-202f55f55e8a" containerID="820f22ba7587e552fc9a6c45c3bc95b62568b98ccf0bd9a89dfd34b991e6d158" exitCode=0 Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.543765 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4hg" event={"ID":"c982d336-6845-4efb-8a7b-202f55f55e8a","Type":"ContainerDied","Data":"820f22ba7587e552fc9a6c45c3bc95b62568b98ccf0bd9a89dfd34b991e6d158"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.543796 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4hg" event={"ID":"c982d336-6845-4efb-8a7b-202f55f55e8a","Type":"ContainerStarted","Data":"3aec9a2a946dd7c2fe00b5994e4cc385a346c54b5d7a10e1b691eb207868c306"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.549627 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lq6m" event={"ID":"62d5d2de-567c-4077-8d82-32833bfb6129","Type":"ContainerStarted","Data":"d13a26f4043f3943b970c54d5bae6b00d4851c5208c7063228a1ea7d54ab4d0e"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.553701 4983 generic.go:334] "Generic (PLEG): container finished" podID="0ccb3b8a-516c-4c88-8350-21a238b80504" containerID="1a58e4ecad95a83c38db98cf957bb9796479c97896fc07969ab0bab68214214f" exitCode=0 Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.553753 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fcr4m" event={"ID":"0ccb3b8a-516c-4c88-8350-21a238b80504","Type":"ContainerDied","Data":"1a58e4ecad95a83c38db98cf957bb9796479c97896fc07969ab0bab68214214f"} Oct 01 09:13:57 crc kubenswrapper[4983]: I1001 09:13:57.596268 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8lq6m" podStartSLOduration=2.075965601 podStartE2EDuration="3.596249395s" podCreationTimestamp="2025-10-01 09:13:54 +0000 UTC" firstStartedPulling="2025-10-01 09:13:55.522904004 +0000 UTC m=+263.512132801" lastFinishedPulling="2025-10-01 09:13:57.043187798 +0000 UTC m=+265.032416595" observedRunningTime="2025-10-01 09:13:57.595664917 +0000 UTC m=+265.584893734" watchObservedRunningTime="2025-10-01 09:13:57.596249395 +0000 UTC m=+265.585478192" Oct 01 09:13:58 crc kubenswrapper[4983]: I1001 09:13:58.562000 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4f8xr" event={"ID":"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63","Type":"ContainerStarted","Data":"f19cc33230596f9b51560597a3fba82f71fd3bca4eefa3628117a9502096a612"} Oct 01 09:13:58 crc kubenswrapper[4983]: I1001 09:13:58.564754 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4hg" event={"ID":"c982d336-6845-4efb-8a7b-202f55f55e8a","Type":"ContainerStarted","Data":"c10b3ed6356857d73f6e1bb9ad7cf412c76dbb08993da15808a65a08741dbc4c"} Oct 01 09:13:58 crc kubenswrapper[4983]: I1001 09:13:58.567715 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fcr4m" event={"ID":"0ccb3b8a-516c-4c88-8350-21a238b80504","Type":"ContainerStarted","Data":"dd49e3ed0a05387c447b78a5723c3c1f0eb05f971732addfde078cf6460172de"} Oct 01 09:13:58 crc kubenswrapper[4983]: I1001 09:13:58.592968 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fcr4m" podStartSLOduration=2.018433372 podStartE2EDuration="4.592944941s" podCreationTimestamp="2025-10-01 09:13:54 +0000 UTC" firstStartedPulling="2025-10-01 09:13:55.526000426 +0000 UTC m=+263.515229223" lastFinishedPulling="2025-10-01 09:13:58.100511995 +0000 UTC m=+266.089740792" observedRunningTime="2025-10-01 09:13:58.591391054 +0000 UTC m=+266.580619851" watchObservedRunningTime="2025-10-01 09:13:58.592944941 +0000 UTC m=+266.582173748" Oct 01 09:13:59 crc kubenswrapper[4983]: I1001 09:13:59.575276 4983 generic.go:334] "Generic (PLEG): container finished" podID="b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63" containerID="f19cc33230596f9b51560597a3fba82f71fd3bca4eefa3628117a9502096a612" exitCode=0 Oct 01 09:13:59 crc kubenswrapper[4983]: I1001 09:13:59.575361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4f8xr" event={"ID":"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63","Type":"ContainerDied","Data":"f19cc33230596f9b51560597a3fba82f71fd3bca4eefa3628117a9502096a612"} Oct 01 09:13:59 crc kubenswrapper[4983]: I1001 09:13:59.577573 4983 generic.go:334] "Generic (PLEG): container finished" podID="c982d336-6845-4efb-8a7b-202f55f55e8a" containerID="c10b3ed6356857d73f6e1bb9ad7cf412c76dbb08993da15808a65a08741dbc4c" exitCode=0 Oct 01 09:13:59 crc kubenswrapper[4983]: I1001 09:13:59.578104 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4hg" event={"ID":"c982d336-6845-4efb-8a7b-202f55f55e8a","Type":"ContainerDied","Data":"c10b3ed6356857d73f6e1bb9ad7cf412c76dbb08993da15808a65a08741dbc4c"} Oct 01 09:14:01 crc kubenswrapper[4983]: I1001 09:14:01.609471 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4f8xr" event={"ID":"b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63","Type":"ContainerStarted","Data":"2f203b3c0d9855aa0cab1cf05842c2f0f17b71abdda1e364ecc07ef10937e907"} Oct 01 09:14:01 crc kubenswrapper[4983]: I1001 09:14:01.612612 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cg4hg" event={"ID":"c982d336-6845-4efb-8a7b-202f55f55e8a","Type":"ContainerStarted","Data":"8ae27995ce7000a5ed21e540027d4e42fba11f85e6e9a0faaf0e1326dd9205b5"} Oct 01 09:14:01 crc kubenswrapper[4983]: I1001 09:14:01.632623 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4f8xr" podStartSLOduration=2.875850889 podStartE2EDuration="5.63260262s" podCreationTimestamp="2025-10-01 09:13:56 +0000 UTC" firstStartedPulling="2025-10-01 09:13:57.543627753 +0000 UTC m=+265.532856550" lastFinishedPulling="2025-10-01 09:14:00.300379484 +0000 UTC m=+268.289608281" observedRunningTime="2025-10-01 09:14:01.629885729 +0000 UTC m=+269.619114556" watchObservedRunningTime="2025-10-01 09:14:01.63260262 +0000 UTC m=+269.621831417" Oct 01 09:14:01 crc kubenswrapper[4983]: I1001 09:14:01.650859 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cg4hg" podStartSLOduration=3.009689707 podStartE2EDuration="5.650843965s" podCreationTimestamp="2025-10-01 09:13:56 +0000 UTC" firstStartedPulling="2025-10-01 09:13:57.546120358 +0000 UTC m=+265.535349175" lastFinishedPulling="2025-10-01 09:14:00.187274646 +0000 UTC m=+268.176503433" observedRunningTime="2025-10-01 09:14:01.64830117 +0000 UTC m=+269.637529957" watchObservedRunningTime="2025-10-01 09:14:01.650843965 +0000 UTC m=+269.640072762" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.422110 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.422497 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.468284 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.624516 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.624580 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.661285 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:14:04 crc kubenswrapper[4983]: I1001 09:14:04.673009 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8lq6m" Oct 01 09:14:05 crc kubenswrapper[4983]: I1001 09:14:05.680541 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fcr4m" Oct 01 09:14:06 crc kubenswrapper[4983]: I1001 09:14:06.843644 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:14:06 crc kubenswrapper[4983]: I1001 09:14:06.843705 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:14:06 crc kubenswrapper[4983]: I1001 09:14:06.887082 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:14:07 crc kubenswrapper[4983]: I1001 09:14:07.007192 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:14:07 crc kubenswrapper[4983]: I1001 09:14:07.007261 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:14:07 crc kubenswrapper[4983]: I1001 09:14:07.049093 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:14:07 crc kubenswrapper[4983]: I1001 09:14:07.693030 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cg4hg" Oct 01 09:14:07 crc kubenswrapper[4983]: I1001 09:14:07.701921 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4f8xr" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.136135 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd"] Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.137753 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.140058 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.140224 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.144942 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd"] Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.233353 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.233474 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.233572 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q559g\" (UniqueName: \"kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.335011 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.335089 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.335146 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q559g\" (UniqueName: \"kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.336510 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.342787 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.354765 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q559g\" (UniqueName: \"kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g\") pod \"collect-profiles-29321835-4dngd\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.465994 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.861579 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd"] Oct 01 09:15:00 crc kubenswrapper[4983]: I1001 09:15:00.967522 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" event={"ID":"d3efb768-769f-4584-918a-cd5a98edeaa8","Type":"ContainerStarted","Data":"3286b2fa40d560f2a93b3b4de0272425ab2d8286f8e8407bae61f9197959bf82"} Oct 01 09:15:01 crc kubenswrapper[4983]: I1001 09:15:01.975494 4983 generic.go:334] "Generic (PLEG): container finished" podID="d3efb768-769f-4584-918a-cd5a98edeaa8" containerID="7068ea29e8606dbc6804ac29ff787ba09e2e3c2018fe0a9f091cbdd84c1550ce" exitCode=0 Oct 01 09:15:01 crc kubenswrapper[4983]: I1001 09:15:01.975556 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" event={"ID":"d3efb768-769f-4584-918a-cd5a98edeaa8","Type":"ContainerDied","Data":"7068ea29e8606dbc6804ac29ff787ba09e2e3c2018fe0a9f091cbdd84c1550ce"} Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.198552 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.272608 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume\") pod \"d3efb768-769f-4584-918a-cd5a98edeaa8\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.273231 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume" (OuterVolumeSpecName: "config-volume") pod "d3efb768-769f-4584-918a-cd5a98edeaa8" (UID: "d3efb768-769f-4584-918a-cd5a98edeaa8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.273328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q559g\" (UniqueName: \"kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g\") pod \"d3efb768-769f-4584-918a-cd5a98edeaa8\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.274017 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume\") pod \"d3efb768-769f-4584-918a-cd5a98edeaa8\" (UID: \"d3efb768-769f-4584-918a-cd5a98edeaa8\") " Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.274218 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3efb768-769f-4584-918a-cd5a98edeaa8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.279421 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g" (OuterVolumeSpecName: "kube-api-access-q559g") pod "d3efb768-769f-4584-918a-cd5a98edeaa8" (UID: "d3efb768-769f-4584-918a-cd5a98edeaa8"). InnerVolumeSpecName "kube-api-access-q559g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.280877 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d3efb768-769f-4584-918a-cd5a98edeaa8" (UID: "d3efb768-769f-4584-918a-cd5a98edeaa8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.375921 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q559g\" (UniqueName: \"kubernetes.io/projected/d3efb768-769f-4584-918a-cd5a98edeaa8-kube-api-access-q559g\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.375970 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d3efb768-769f-4584-918a-cd5a98edeaa8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.999261 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" event={"ID":"d3efb768-769f-4584-918a-cd5a98edeaa8","Type":"ContainerDied","Data":"3286b2fa40d560f2a93b3b4de0272425ab2d8286f8e8407bae61f9197959bf82"} Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.999318 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3286b2fa40d560f2a93b3b4de0272425ab2d8286f8e8407bae61f9197959bf82" Oct 01 09:15:03 crc kubenswrapper[4983]: I1001 09:15:03.999331 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-4dngd" Oct 01 09:15:31 crc kubenswrapper[4983]: I1001 09:15:31.960456 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:15:31 crc kubenswrapper[4983]: I1001 09:15:31.961075 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:16:01 crc kubenswrapper[4983]: I1001 09:16:01.961041 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:16:01 crc kubenswrapper[4983]: I1001 09:16:01.961422 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:16:31 crc kubenswrapper[4983]: I1001 09:16:31.961217 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:16:31 crc kubenswrapper[4983]: I1001 09:16:31.962054 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:16:31 crc kubenswrapper[4983]: I1001 09:16:31.962126 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:16:31 crc kubenswrapper[4983]: I1001 09:16:31.962958 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:16:31 crc kubenswrapper[4983]: I1001 09:16:31.963053 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3" gracePeriod=600 Oct 01 09:16:32 crc kubenswrapper[4983]: I1001 09:16:32.530740 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3" exitCode=0 Oct 01 09:16:32 crc kubenswrapper[4983]: I1001 09:16:32.530853 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3"} Oct 01 09:16:32 crc kubenswrapper[4983]: I1001 09:16:32.531407 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533"} Oct 01 09:16:32 crc kubenswrapper[4983]: I1001 09:16:32.531435 4983 scope.go:117] "RemoveContainer" containerID="b623e9ea9ff4e4cd5ad651288753f3609e609462d2bb568c91bea90b01d5d198" Oct 01 09:17:30 crc kubenswrapper[4983]: I1001 09:17:30.902062 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttrjq"] Oct 01 09:17:30 crc kubenswrapper[4983]: E1001 09:17:30.902999 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3efb768-769f-4584-918a-cd5a98edeaa8" containerName="collect-profiles" Oct 01 09:17:30 crc kubenswrapper[4983]: I1001 09:17:30.903015 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3efb768-769f-4584-918a-cd5a98edeaa8" containerName="collect-profiles" Oct 01 09:17:30 crc kubenswrapper[4983]: I1001 09:17:30.903119 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3efb768-769f-4584-918a-cd5a98edeaa8" containerName="collect-profiles" Oct 01 09:17:30 crc kubenswrapper[4983]: I1001 09:17:30.903578 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:30 crc kubenswrapper[4983]: I1001 09:17:30.911032 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttrjq"] Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029146 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029207 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029247 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-bound-sa-token\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029262 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-certificates\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029278 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57znw\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-kube-api-access-57znw\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029302 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029325 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-tls\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.029406 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-trusted-ca\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.048079 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.130989 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131092 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-bound-sa-token\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131129 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-certificates\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131161 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57znw\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-kube-api-access-57znw\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131217 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131265 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-tls\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.131309 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-trusted-ca\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.132143 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.132802 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-certificates\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.133710 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-trusted-ca\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.137206 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-registry-tls\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.137669 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.148056 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-bound-sa-token\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.157940 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57znw\" (UniqueName: \"kubernetes.io/projected/912a9eb7-96d2-4c8c-83d5-ef291e3859c0-kube-api-access-57znw\") pod \"image-registry-66df7c8f76-ttrjq\" (UID: \"912a9eb7-96d2-4c8c-83d5-ef291e3859c0\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.219358 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.445417 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttrjq"] Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.875831 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" event={"ID":"912a9eb7-96d2-4c8c-83d5-ef291e3859c0","Type":"ContainerStarted","Data":"3cba3779f026daf4e9002bc0eec4808ae2320a8c7a47ad21cea6c9caddb06a17"} Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.876105 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" event={"ID":"912a9eb7-96d2-4c8c-83d5-ef291e3859c0","Type":"ContainerStarted","Data":"b67d3b3eb81acf72a8af528b7b351c73366e40e9b901ddf265b2199ff1b547ac"} Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.876140 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:31 crc kubenswrapper[4983]: I1001 09:17:31.899427 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" podStartSLOduration=1.899399842 podStartE2EDuration="1.899399842s" podCreationTimestamp="2025-10-01 09:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:17:31.895125627 +0000 UTC m=+479.884354514" watchObservedRunningTime="2025-10-01 09:17:31.899399842 +0000 UTC m=+479.888628679" Oct 01 09:17:51 crc kubenswrapper[4983]: I1001 09:17:51.225446 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-ttrjq" Oct 01 09:17:51 crc kubenswrapper[4983]: I1001 09:17:51.278476 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.322286 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" podUID="bde163a6-2b23-4f0f-873b-a37d2213d232" containerName="registry" containerID="cri-o://bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29" gracePeriod=30 Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.716396 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.813264 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv9cs\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.813516 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.813838 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.813946 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.814005 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.814064 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.814144 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.814178 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token\") pod \"bde163a6-2b23-4f0f-873b-a37d2213d232\" (UID: \"bde163a6-2b23-4f0f-873b-a37d2213d232\") " Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.814966 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.816710 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.820001 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.820058 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs" (OuterVolumeSpecName: "kube-api-access-rv9cs") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "kube-api-access-rv9cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.820118 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.820900 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.827275 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.834644 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bde163a6-2b23-4f0f-873b-a37d2213d232" (UID: "bde163a6-2b23-4f0f-873b-a37d2213d232"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915513 4983 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915552 4983 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915569 4983 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bde163a6-2b23-4f0f-873b-a37d2213d232-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915580 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bde163a6-2b23-4f0f-873b-a37d2213d232-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915591 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915601 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv9cs\" (UniqueName: \"kubernetes.io/projected/bde163a6-2b23-4f0f-873b-a37d2213d232-kube-api-access-rv9cs\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:16 crc kubenswrapper[4983]: I1001 09:18:16.915614 4983 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bde163a6-2b23-4f0f-873b-a37d2213d232-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.136557 4983 generic.go:334] "Generic (PLEG): container finished" podID="bde163a6-2b23-4f0f-873b-a37d2213d232" containerID="bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29" exitCode=0 Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.136626 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.136629 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" event={"ID":"bde163a6-2b23-4f0f-873b-a37d2213d232","Type":"ContainerDied","Data":"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29"} Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.136657 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2pr5s" event={"ID":"bde163a6-2b23-4f0f-873b-a37d2213d232","Type":"ContainerDied","Data":"87cb0ccdb84c5232ccd43c35871c8af02ee61d818379f598de49bfb468bb92b5"} Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.136677 4983 scope.go:117] "RemoveContainer" containerID="bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29" Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.154700 4983 scope.go:117] "RemoveContainer" containerID="bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29" Oct 01 09:18:17 crc kubenswrapper[4983]: E1001 09:18:17.155231 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29\": container with ID starting with bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29 not found: ID does not exist" containerID="bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29" Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.155280 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29"} err="failed to get container status \"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29\": rpc error: code = NotFound desc = could not find container \"bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29\": container with ID starting with bb1ae919435177cae2c93f35064098e0f473f8a860bd8804db62803fccdeda29 not found: ID does not exist" Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.163857 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:18:17 crc kubenswrapper[4983]: I1001 09:18:17.167266 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2pr5s"] Oct 01 09:18:18 crc kubenswrapper[4983]: I1001 09:18:18.722724 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde163a6-2b23-4f0f-873b-a37d2213d232" path="/var/lib/kubelet/pods/bde163a6-2b23-4f0f-873b-a37d2213d232/volumes" Oct 01 09:19:01 crc kubenswrapper[4983]: I1001 09:19:01.961009 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:19:01 crc kubenswrapper[4983]: I1001 09:19:01.961586 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.620148 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7skk"] Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621133 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-controller" containerID="cri-o://a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621206 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="nbdb" containerID="cri-o://72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621274 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-acl-logging" containerID="cri-o://c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621276 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621260 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-node" containerID="cri-o://5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621408 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="sbdb" containerID="cri-o://9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.621508 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="northd" containerID="cri-o://0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.661183 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" containerID="cri-o://edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" gracePeriod=30 Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.898624 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/3.log" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.900969 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovn-acl-logging/0.log" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.901519 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovn-controller/0.log" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.901973 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950353 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2zmzc"] Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950558 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-acl-logging" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950575 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-acl-logging" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950612 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="northd" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950620 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="northd" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950630 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="sbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950639 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="sbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950650 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kubecfg-setup" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950658 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kubecfg-setup" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950670 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950677 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950687 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde163a6-2b23-4f0f-873b-a37d2213d232" containerName="registry" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950695 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde163a6-2b23-4f0f-873b-a37d2213d232" containerName="registry" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950721 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950729 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950737 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950745 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950754 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-node" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950761 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-node" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950779 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950786 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950796 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950803 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950869 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="nbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950876 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="nbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.950885 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.950892 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951079 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951095 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951104 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="sbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951111 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951121 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="northd" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951129 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="nbdb" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951137 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951145 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde163a6-2b23-4f0f-873b-a37d2213d232" containerName="registry" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951153 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951165 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovn-acl-logging" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951175 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951184 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="kube-rbac-proxy-node" Oct 01 09:19:21 crc kubenswrapper[4983]: E1001 09:19:21.951270 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951277 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.951365 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerName="ovnkube-controller" Oct 01 09:19:21 crc kubenswrapper[4983]: I1001 09:19:21.953050 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030395 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030607 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030667 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030682 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030731 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030758 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030823 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031255 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031350 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031378 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031461 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031477 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031499 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031514 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031530 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031556 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031586 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpzmm\" (UniqueName: \"kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030889 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.030961 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031623 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031699 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031719 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031739 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash\") pod \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\" (UID: \"3e374b0f-544c-458e-9de6-5dd3149c3dd0\") " Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031653 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031019 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031186 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031308 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log" (OuterVolumeSpecName: "node-log") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031412 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031608 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031616 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031675 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket" (OuterVolumeSpecName: "log-socket") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031694 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031711 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031738 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031981 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash" (OuterVolumeSpecName: "host-slash") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032019 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032137 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032150 4983 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032159 4983 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032167 4983 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032177 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032186 4983 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-node-log\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032194 4983 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032203 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e374b0f-544c-458e-9de6-5dd3149c3dd0-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032211 4983 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032219 4983 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032227 4983 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032235 4983 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032243 4983 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-log-socket\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032251 4983 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032260 4983 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.032267 4983 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-slash\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.031947 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.036081 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm" (OuterVolumeSpecName: "kube-api-access-lpzmm") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "kube-api-access-lpzmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.036241 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.042850 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3e374b0f-544c-458e-9de6-5dd3149c3dd0" (UID: "3e374b0f-544c-458e-9de6-5dd3149c3dd0"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133464 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-ovn\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133543 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-systemd-units\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133574 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133601 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133625 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wngsz\" (UniqueName: \"kubernetes.io/projected/59841016-319f-461d-865d-a7db9a56a867-kube-api-access-wngsz\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133649 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-log-socket\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133673 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-var-lib-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133695 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133721 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-config\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133743 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-slash\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133767 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-kubelet\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133793 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-netns\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133837 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-netd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133895 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-systemd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133917 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-etc-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133936 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-script-lib\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.133999 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-node-log\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134020 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59841016-319f-461d-865d-a7db9a56a867-ovn-node-metrics-cert\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134040 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-env-overrides\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-bin\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134105 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpzmm\" (UniqueName: \"kubernetes.io/projected/3e374b0f-544c-458e-9de6-5dd3149c3dd0-kube-api-access-lpzmm\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134116 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e374b0f-544c-458e-9de6-5dd3149c3dd0-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134125 4983 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.134133 4983 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e374b0f-544c-458e-9de6-5dd3149c3dd0-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.235485 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-node-log\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.235926 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59841016-319f-461d-865d-a7db9a56a867-ovn-node-metrics-cert\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.235970 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-env-overrides\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236006 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-bin\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.235603 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-node-log\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236044 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-ovn\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236096 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-systemd-units\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236126 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-systemd-units\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236125 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-ovn\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236096 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-bin\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236134 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236193 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236216 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236198 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-ovn-kubernetes\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236265 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wngsz\" (UniqueName: \"kubernetes.io/projected/59841016-319f-461d-865d-a7db9a56a867-kube-api-access-wngsz\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236309 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-log-socket\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236344 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-var-lib-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236380 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236414 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-config\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236425 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-log-socket\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236445 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-var-lib-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236445 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-slash\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-slash\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236492 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-kubelet\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236497 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236512 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-kubelet\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236536 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-netns\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236574 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-netd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236611 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-systemd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236644 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-script-lib\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236652 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-cni-netd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236676 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-etc-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236684 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-run-systemd\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236649 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-host-run-netns\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236653 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-env-overrides\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236727 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59841016-319f-461d-865d-a7db9a56a867-etc-openvswitch\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.236955 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-config\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.237271 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59841016-319f-461d-865d-a7db9a56a867-ovnkube-script-lib\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.239698 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59841016-319f-461d-865d-a7db9a56a867-ovn-node-metrics-cert\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.252205 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wngsz\" (UniqueName: \"kubernetes.io/projected/59841016-319f-461d-865d-a7db9a56a867-kube-api-access-wngsz\") pod \"ovnkube-node-2zmzc\" (UID: \"59841016-319f-461d-865d-a7db9a56a867\") " pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.264097 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.502270 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovnkube-controller/3.log" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.504781 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovn-acl-logging/0.log" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505218 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7skk_3e374b0f-544c-458e-9de6-5dd3149c3dd0/ovn-controller/0.log" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505647 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505686 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505698 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505707 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505717 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505730 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505739 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" exitCode=143 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505749 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" exitCode=143 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505760 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505777 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505858 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505838 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505971 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.505994 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506008 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506022 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506036 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506049 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506057 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506064 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506072 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506082 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506090 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506096 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506103 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506127 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506136 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506142 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506150 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506157 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506164 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506172 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506179 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506186 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506193 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506203 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506215 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506224 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506230 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506237 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506247 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506254 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506261 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506268 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506274 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506282 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506291 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7skk" event={"ID":"3e374b0f-544c-458e-9de6-5dd3149c3dd0","Type":"ContainerDied","Data":"c567ee932754affeaa92f13ca0fbdf29c244b3ce5e03efb966aa7c42d67333ae"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506302 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506310 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506317 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506324 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506330 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506339 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506346 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506353 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506359 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.506365 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.507709 4983 generic.go:334] "Generic (PLEG): container finished" podID="59841016-319f-461d-865d-a7db9a56a867" containerID="fe375ae38b53ed0e2f08fa477446f43d4e1130a4e4e0fde0fe83d4b5be89ce21" exitCode=0 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.507767 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerDied","Data":"fe375ae38b53ed0e2f08fa477446f43d4e1130a4e4e0fde0fe83d4b5be89ce21"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.507794 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"a6412205a83026a1f4b3d3a226bde6ad167cee31c26a867577bb1780afcc3b5a"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.510620 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/2.log" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.511097 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/1.log" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.511125 4983 generic.go:334] "Generic (PLEG): container finished" podID="eb6d10ad-935a-496d-8c25-32e024026cd9" containerID="2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662" exitCode=2 Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.511142 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerDied","Data":"2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.511157 4983 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01"} Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.511383 4983 scope.go:117] "RemoveContainer" containerID="2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.511523 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-p92ll_openshift-multus(eb6d10ad-935a-496d-8c25-32e024026cd9)\"" pod="openshift-multus/multus-p92ll" podUID="eb6d10ad-935a-496d-8c25-32e024026cd9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.527138 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.545544 4983 scope.go:117] "RemoveContainer" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.564885 4983 scope.go:117] "RemoveContainer" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.572497 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7skk"] Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.578302 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7skk"] Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.587600 4983 scope.go:117] "RemoveContainer" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.600095 4983 scope.go:117] "RemoveContainer" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.612270 4983 scope.go:117] "RemoveContainer" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.630840 4983 scope.go:117] "RemoveContainer" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.641109 4983 scope.go:117] "RemoveContainer" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.655921 4983 scope.go:117] "RemoveContainer" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.674691 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.675077 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.675120 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} err="failed to get container status \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.675150 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.675592 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": container with ID starting with ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782 not found: ID does not exist" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.675625 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} err="failed to get container status \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": rpc error: code = NotFound desc = could not find container \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": container with ID starting with ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.675644 4983 scope.go:117] "RemoveContainer" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.676010 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": container with ID starting with 9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845 not found: ID does not exist" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.676079 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} err="failed to get container status \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": rpc error: code = NotFound desc = could not find container \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": container with ID starting with 9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.676136 4983 scope.go:117] "RemoveContainer" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.676854 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": container with ID starting with 72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6 not found: ID does not exist" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.676890 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} err="failed to get container status \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": rpc error: code = NotFound desc = could not find container \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": container with ID starting with 72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.676909 4983 scope.go:117] "RemoveContainer" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.677210 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": container with ID starting with 0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4 not found: ID does not exist" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.677232 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} err="failed to get container status \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": rpc error: code = NotFound desc = could not find container \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": container with ID starting with 0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.677246 4983 scope.go:117] "RemoveContainer" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.677751 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": container with ID starting with 626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f not found: ID does not exist" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.677836 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} err="failed to get container status \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": rpc error: code = NotFound desc = could not find container \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": container with ID starting with 626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.677870 4983 scope.go:117] "RemoveContainer" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.678952 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": container with ID starting with 5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5 not found: ID does not exist" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.678977 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} err="failed to get container status \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": rpc error: code = NotFound desc = could not find container \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": container with ID starting with 5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.678995 4983 scope.go:117] "RemoveContainer" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.679590 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": container with ID starting with c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9 not found: ID does not exist" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.679635 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} err="failed to get container status \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": rpc error: code = NotFound desc = could not find container \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": container with ID starting with c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.679656 4983 scope.go:117] "RemoveContainer" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.679961 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": container with ID starting with a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c not found: ID does not exist" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680007 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} err="failed to get container status \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": rpc error: code = NotFound desc = could not find container \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": container with ID starting with a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680023 4983 scope.go:117] "RemoveContainer" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: E1001 09:19:22.680437 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": container with ID starting with 1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b not found: ID does not exist" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680483 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} err="failed to get container status \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": rpc error: code = NotFound desc = could not find container \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": container with ID starting with 1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680499 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680817 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} err="failed to get container status \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.680852 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681154 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} err="failed to get container status \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": rpc error: code = NotFound desc = could not find container \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": container with ID starting with ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681176 4983 scope.go:117] "RemoveContainer" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681474 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} err="failed to get container status \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": rpc error: code = NotFound desc = could not find container \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": container with ID starting with 9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681497 4983 scope.go:117] "RemoveContainer" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681910 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} err="failed to get container status \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": rpc error: code = NotFound desc = could not find container \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": container with ID starting with 72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.681930 4983 scope.go:117] "RemoveContainer" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.682941 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} err="failed to get container status \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": rpc error: code = NotFound desc = could not find container \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": container with ID starting with 0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.682961 4983 scope.go:117] "RemoveContainer" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.683380 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} err="failed to get container status \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": rpc error: code = NotFound desc = could not find container \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": container with ID starting with 626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.683404 4983 scope.go:117] "RemoveContainer" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.683680 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} err="failed to get container status \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": rpc error: code = NotFound desc = could not find container \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": container with ID starting with 5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.683698 4983 scope.go:117] "RemoveContainer" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684185 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} err="failed to get container status \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": rpc error: code = NotFound desc = could not find container \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": container with ID starting with c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684202 4983 scope.go:117] "RemoveContainer" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684541 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} err="failed to get container status \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": rpc error: code = NotFound desc = could not find container \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": container with ID starting with a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684562 4983 scope.go:117] "RemoveContainer" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684885 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} err="failed to get container status \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": rpc error: code = NotFound desc = could not find container \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": container with ID starting with 1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.684905 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685129 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} err="failed to get container status \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685147 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685360 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} err="failed to get container status \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": rpc error: code = NotFound desc = could not find container \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": container with ID starting with ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685382 4983 scope.go:117] "RemoveContainer" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685610 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} err="failed to get container status \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": rpc error: code = NotFound desc = could not find container \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": container with ID starting with 9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685628 4983 scope.go:117] "RemoveContainer" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685843 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} err="failed to get container status \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": rpc error: code = NotFound desc = could not find container \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": container with ID starting with 72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.685870 4983 scope.go:117] "RemoveContainer" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686194 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} err="failed to get container status \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": rpc error: code = NotFound desc = could not find container \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": container with ID starting with 0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686216 4983 scope.go:117] "RemoveContainer" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686561 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} err="failed to get container status \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": rpc error: code = NotFound desc = could not find container \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": container with ID starting with 626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686581 4983 scope.go:117] "RemoveContainer" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686934 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} err="failed to get container status \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": rpc error: code = NotFound desc = could not find container \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": container with ID starting with 5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.686962 4983 scope.go:117] "RemoveContainer" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687196 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} err="failed to get container status \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": rpc error: code = NotFound desc = could not find container \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": container with ID starting with c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687218 4983 scope.go:117] "RemoveContainer" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687462 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} err="failed to get container status \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": rpc error: code = NotFound desc = could not find container \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": container with ID starting with a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687484 4983 scope.go:117] "RemoveContainer" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687795 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} err="failed to get container status \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": rpc error: code = NotFound desc = could not find container \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": container with ID starting with 1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.687831 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688084 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} err="failed to get container status \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688103 4983 scope.go:117] "RemoveContainer" containerID="ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688406 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782"} err="failed to get container status \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": rpc error: code = NotFound desc = could not find container \"ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782\": container with ID starting with ccf61e86c6142432f52940c92321b96922f3940fc285604aaba39a94cb312782 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688425 4983 scope.go:117] "RemoveContainer" containerID="9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688649 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845"} err="failed to get container status \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": rpc error: code = NotFound desc = could not find container \"9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845\": container with ID starting with 9958a5dbf42e69d973acb06da2fc22037ccbb64fa8d63731d3275d6ff14a5845 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688672 4983 scope.go:117] "RemoveContainer" containerID="72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688975 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6"} err="failed to get container status \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": rpc error: code = NotFound desc = could not find container \"72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6\": container with ID starting with 72d4d1f154e08d4da2ddc6d85054e97574b04674d5399002c90278946bedd3b6 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.688997 4983 scope.go:117] "RemoveContainer" containerID="0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689300 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4"} err="failed to get container status \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": rpc error: code = NotFound desc = could not find container \"0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4\": container with ID starting with 0f8bca2c0f1ecb0eced3380592d894fbf6fef19797c3fa9cd4db864b8bb22bb4 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689317 4983 scope.go:117] "RemoveContainer" containerID="626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689521 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f"} err="failed to get container status \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": rpc error: code = NotFound desc = could not find container \"626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f\": container with ID starting with 626b4f737d6ceca2df2bf83f932788107e05efb8e00b9e0bc89aef19f8d4317f not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689548 4983 scope.go:117] "RemoveContainer" containerID="5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689753 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5"} err="failed to get container status \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": rpc error: code = NotFound desc = could not find container \"5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5\": container with ID starting with 5f5a90de94336c2bdc1c92b5dab880317a91f99382e2c95a41902037f562a5f5 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.689774 4983 scope.go:117] "RemoveContainer" containerID="c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690019 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9"} err="failed to get container status \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": rpc error: code = NotFound desc = could not find container \"c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9\": container with ID starting with c829bc9744a5c22f752c4a886f76de4e998a33c2e03cd1f91f4b35022a3836c9 not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690045 4983 scope.go:117] "RemoveContainer" containerID="a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690323 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c"} err="failed to get container status \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": rpc error: code = NotFound desc = could not find container \"a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c\": container with ID starting with a02f4d3d6ad5cf311b3ff6ed6cfac05f9484d58e55a8e1d8b15e8e9dcf46a83c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690347 4983 scope.go:117] "RemoveContainer" containerID="1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690567 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b"} err="failed to get container status \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": rpc error: code = NotFound desc = could not find container \"1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b\": container with ID starting with 1c8c034166e3a3da46b4f5e9ac25f55ba085477be767e4f95e01d4a86c35247b not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690588 4983 scope.go:117] "RemoveContainer" containerID="edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.690864 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c"} err="failed to get container status \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": rpc error: code = NotFound desc = could not find container \"edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c\": container with ID starting with edb7091d22c91267625fad3e69cd45b169076a7a3b6e9cea2fe1f5d51b9cbd3c not found: ID does not exist" Oct 01 09:19:22 crc kubenswrapper[4983]: I1001 09:19:22.721932 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e374b0f-544c-458e-9de6-5dd3149c3dd0" path="/var/lib/kubelet/pods/3e374b0f-544c-458e-9de6-5dd3149c3dd0/volumes" Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.522887 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"7befb33dbbf288119f175b31fc34b89482b3775cf6e907e36430ee1bf227af29"} Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.523274 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"1084cab879434ea12ce510f8cae6128d51aa21f8c23024596626e13d4e090706"} Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.523295 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"21d034ed1cceb810892dfe2cc40c6841672a0080e70bd6ead025b47a0d042fa1"} Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.523315 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"21c72a9dc5c695c4fa37fd8786fea8cd285080e3dfd0c852a32b2153e557972b"} Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.523332 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"f4fbf1399e702843af165a5520daa7038488f05a48297f4dd51c365f3313b607"} Oct 01 09:19:23 crc kubenswrapper[4983]: I1001 09:19:23.523348 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"21b0a56cc003d21b251dbd27142a05644e142571f535b2e7474c5ee67eec5d56"} Oct 01 09:19:25 crc kubenswrapper[4983]: I1001 09:19:25.540159 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"b5c8bd87a175b01c569692edafb25fce385aa582f61df8af0bc60809504cb305"} Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.561986 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" event={"ID":"59841016-319f-461d-865d-a7db9a56a867","Type":"ContainerStarted","Data":"5cbe5d748e0474a1a78da2455449e7042c59de35d7b9aa1873ddb37abe1dbbfa"} Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.562640 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.562662 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.562673 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.586239 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" podStartSLOduration=7.586223089 podStartE2EDuration="7.586223089s" podCreationTimestamp="2025-10-01 09:19:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:19:28.585978603 +0000 UTC m=+596.575207410" watchObservedRunningTime="2025-10-01 09:19:28.586223089 +0000 UTC m=+596.575451886" Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.594334 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:28 crc kubenswrapper[4983]: I1001 09:19:28.594689 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:31 crc kubenswrapper[4983]: I1001 09:19:31.960363 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:19:31 crc kubenswrapper[4983]: I1001 09:19:31.960599 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:19:32 crc kubenswrapper[4983]: I1001 09:19:32.884669 4983 scope.go:117] "RemoveContainer" containerID="4dd20158bcb10bca702b9d28b258c911b7e8ac3d1304fea847f3ff08c0ee8e01" Oct 01 09:19:33 crc kubenswrapper[4983]: I1001 09:19:33.590908 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/2.log" Oct 01 09:19:36 crc kubenswrapper[4983]: I1001 09:19:36.715784 4983 scope.go:117] "RemoveContainer" containerID="2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662" Oct 01 09:19:36 crc kubenswrapper[4983]: E1001 09:19:36.716666 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-p92ll_openshift-multus(eb6d10ad-935a-496d-8c25-32e024026cd9)\"" pod="openshift-multus/multus-p92ll" podUID="eb6d10ad-935a-496d-8c25-32e024026cd9" Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.793607 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6"] Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.796861 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.801747 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.809763 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6"] Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.919331 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.919399 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:45 crc kubenswrapper[4983]: I1001 09:19:45.919474 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhmbn\" (UniqueName: \"kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.020594 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhmbn\" (UniqueName: \"kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.020786 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.020920 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.021534 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.021682 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.053088 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhmbn\" (UniqueName: \"kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.121390 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.154053 4983 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(93fa03d71d02089016956c548d89642a8756eb5ff3b080a2c7f038848f9342db): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.154178 4983 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(93fa03d71d02089016956c548d89642a8756eb5ff3b080a2c7f038848f9342db): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.154204 4983 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(93fa03d71d02089016956c548d89642a8756eb5ff3b080a2c7f038848f9342db): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.154250 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace(a7fe723e-4359-4361-ba69-cb6411ffbbf3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace(a7fe723e-4359-4361-ba69-cb6411ffbbf3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(93fa03d71d02089016956c548d89642a8756eb5ff3b080a2c7f038848f9342db): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.662850 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: I1001 09:19:46.663591 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.708256 4983 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(36f05cc60c70c9ac78b44ae01d15e1ddf24036ec258d18a7f8228a5c8f844d6e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.708359 4983 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(36f05cc60c70c9ac78b44ae01d15e1ddf24036ec258d18a7f8228a5c8f844d6e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.708382 4983 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(36f05cc60c70c9ac78b44ae01d15e1ddf24036ec258d18a7f8228a5c8f844d6e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:46 crc kubenswrapper[4983]: E1001 09:19:46.708434 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace(a7fe723e-4359-4361-ba69-cb6411ffbbf3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace(a7fe723e-4359-4361-ba69-cb6411ffbbf3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_openshift-marketplace_a7fe723e-4359-4361-ba69-cb6411ffbbf3_0(36f05cc60c70c9ac78b44ae01d15e1ddf24036ec258d18a7f8228a5c8f844d6e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" Oct 01 09:19:50 crc kubenswrapper[4983]: I1001 09:19:50.714448 4983 scope.go:117] "RemoveContainer" containerID="2f73d3ef0891da16948fbb22ae4f3e147b1727869381d8c459f7a10f92ec6662" Oct 01 09:19:51 crc kubenswrapper[4983]: I1001 09:19:51.690028 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p92ll_eb6d10ad-935a-496d-8c25-32e024026cd9/kube-multus/2.log" Oct 01 09:19:51 crc kubenswrapper[4983]: I1001 09:19:51.690421 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p92ll" event={"ID":"eb6d10ad-935a-496d-8c25-32e024026cd9","Type":"ContainerStarted","Data":"2a47bf04f4bfef0c9bb30a4e8317c02f0ba80e5c2c62308f4838e6bc55d49263"} Oct 01 09:19:52 crc kubenswrapper[4983]: I1001 09:19:52.290038 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2zmzc" Oct 01 09:19:59 crc kubenswrapper[4983]: I1001 09:19:59.714061 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:59 crc kubenswrapper[4983]: I1001 09:19:59.715227 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:19:59 crc kubenswrapper[4983]: I1001 09:19:59.960568 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6"] Oct 01 09:19:59 crc kubenswrapper[4983]: W1001 09:19:59.970289 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7fe723e_4359_4361_ba69_cb6411ffbbf3.slice/crio-64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852 WatchSource:0}: Error finding container 64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852: Status 404 returned error can't find the container with id 64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852 Oct 01 09:20:00 crc kubenswrapper[4983]: I1001 09:20:00.743950 4983 generic.go:334] "Generic (PLEG): container finished" podID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerID="19a0f2bf43c45ca8353591ac87d0d7e2d8596258b21b5060961b4b80daff26d0" exitCode=0 Oct 01 09:20:00 crc kubenswrapper[4983]: I1001 09:20:00.744004 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" event={"ID":"a7fe723e-4359-4361-ba69-cb6411ffbbf3","Type":"ContainerDied","Data":"19a0f2bf43c45ca8353591ac87d0d7e2d8596258b21b5060961b4b80daff26d0"} Oct 01 09:20:00 crc kubenswrapper[4983]: I1001 09:20:00.744038 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" event={"ID":"a7fe723e-4359-4361-ba69-cb6411ffbbf3","Type":"ContainerStarted","Data":"64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852"} Oct 01 09:20:00 crc kubenswrapper[4983]: I1001 09:20:00.745742 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:20:01 crc kubenswrapper[4983]: I1001 09:20:01.961111 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:20:01 crc kubenswrapper[4983]: I1001 09:20:01.961171 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:20:01 crc kubenswrapper[4983]: I1001 09:20:01.961214 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:20:01 crc kubenswrapper[4983]: I1001 09:20:01.961649 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:20:01 crc kubenswrapper[4983]: I1001 09:20:01.961742 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533" gracePeriod=600 Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.757486 4983 generic.go:334] "Generic (PLEG): container finished" podID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerID="361546bd57a95fdf49061702bf84822a4b73b6fdafd4e8a36f8fa14c3a0d6399" exitCode=0 Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.757596 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" event={"ID":"a7fe723e-4359-4361-ba69-cb6411ffbbf3","Type":"ContainerDied","Data":"361546bd57a95fdf49061702bf84822a4b73b6fdafd4e8a36f8fa14c3a0d6399"} Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.761023 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533" exitCode=0 Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.761053 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533"} Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.761089 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7"} Oct 01 09:20:02 crc kubenswrapper[4983]: I1001 09:20:02.761106 4983 scope.go:117] "RemoveContainer" containerID="d821434525823e1fd45e4b56f1fb001cb6a3788f34cd7765e8b27c9886a3d3d3" Oct 01 09:20:03 crc kubenswrapper[4983]: I1001 09:20:03.775392 4983 generic.go:334] "Generic (PLEG): container finished" podID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerID="03e3ec53c0665e267f484b0e9308ae04f0e4934e4fb020ccc4a29027fb24fde5" exitCode=0 Oct 01 09:20:03 crc kubenswrapper[4983]: I1001 09:20:03.775746 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" event={"ID":"a7fe723e-4359-4361-ba69-cb6411ffbbf3","Type":"ContainerDied","Data":"03e3ec53c0665e267f484b0e9308ae04f0e4934e4fb020ccc4a29027fb24fde5"} Oct 01 09:20:04 crc kubenswrapper[4983]: I1001 09:20:04.999660 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.156781 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhmbn\" (UniqueName: \"kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn\") pod \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.157141 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle\") pod \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.157212 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util\") pod \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\" (UID: \"a7fe723e-4359-4361-ba69-cb6411ffbbf3\") " Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.158070 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle" (OuterVolumeSpecName: "bundle") pod "a7fe723e-4359-4361-ba69-cb6411ffbbf3" (UID: "a7fe723e-4359-4361-ba69-cb6411ffbbf3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.162482 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn" (OuterVolumeSpecName: "kube-api-access-vhmbn") pod "a7fe723e-4359-4361-ba69-cb6411ffbbf3" (UID: "a7fe723e-4359-4361-ba69-cb6411ffbbf3"). InnerVolumeSpecName "kube-api-access-vhmbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.170544 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util" (OuterVolumeSpecName: "util") pod "a7fe723e-4359-4361-ba69-cb6411ffbbf3" (UID: "a7fe723e-4359-4361-ba69-cb6411ffbbf3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.258932 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.258985 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7fe723e-4359-4361-ba69-cb6411ffbbf3-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.258998 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhmbn\" (UniqueName: \"kubernetes.io/projected/a7fe723e-4359-4361-ba69-cb6411ffbbf3-kube-api-access-vhmbn\") on node \"crc\" DevicePath \"\"" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.790887 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" event={"ID":"a7fe723e-4359-4361-ba69-cb6411ffbbf3","Type":"ContainerDied","Data":"64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852"} Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.790934 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64235a3ea00af40f978f78f0fed0a67234147532d58a45642d575993f3cff852" Oct 01 09:20:05 crc kubenswrapper[4983]: I1001 09:20:05.791066 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.936115 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv"] Oct 01 09:20:18 crc kubenswrapper[4983]: E1001 09:20:18.936987 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="extract" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.937000 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="extract" Oct 01 09:20:18 crc kubenswrapper[4983]: E1001 09:20:18.937020 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="pull" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.937027 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="pull" Oct 01 09:20:18 crc kubenswrapper[4983]: E1001 09:20:18.937045 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="util" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.937051 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="util" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.937255 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7fe723e-4359-4361-ba69-cb6411ffbbf3" containerName="extract" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.937852 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.944894 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.945032 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.945942 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zrv4r" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.946006 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.946044 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 01 09:20:18 crc kubenswrapper[4983]: I1001 09:20:18.959691 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv"] Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.021962 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.022042 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8njv7\" (UniqueName: \"kubernetes.io/projected/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-kube-api-access-8njv7\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.022191 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-webhook-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.123403 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.123460 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8njv7\" (UniqueName: \"kubernetes.io/projected/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-kube-api-access-8njv7\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.123519 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-webhook-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.130485 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-webhook-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.132484 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.149449 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8njv7\" (UniqueName: \"kubernetes.io/projected/31e9843f-42b4-4229-bdfe-d64ccd1b37eb-kube-api-access-8njv7\") pod \"metallb-operator-controller-manager-7d59c47655-9kwhv\" (UID: \"31e9843f-42b4-4229-bdfe-d64ccd1b37eb\") " pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.162397 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-67c678544c-thtq2"] Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.163187 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.165087 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.165121 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.165209 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5bnfm" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.173668 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-67c678544c-thtq2"] Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.266076 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.325644 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-webhook-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.325693 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7trhc\" (UniqueName: \"kubernetes.io/projected/7a7230d0-bbf6-4949-9af0-6489822f46f5-kube-api-access-7trhc\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.325882 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-apiservice-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.427712 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-webhook-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.427969 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7trhc\" (UniqueName: \"kubernetes.io/projected/7a7230d0-bbf6-4949-9af0-6489822f46f5-kube-api-access-7trhc\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.428774 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-apiservice-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.434251 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-webhook-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.435127 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a7230d0-bbf6-4949-9af0-6489822f46f5-apiservice-cert\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.448466 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7trhc\" (UniqueName: \"kubernetes.io/projected/7a7230d0-bbf6-4949-9af0-6489822f46f5-kube-api-access-7trhc\") pod \"metallb-operator-webhook-server-67c678544c-thtq2\" (UID: \"7a7230d0-bbf6-4949-9af0-6489822f46f5\") " pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.478176 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv"] Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.488148 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.725940 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-67c678544c-thtq2"] Oct 01 09:20:19 crc kubenswrapper[4983]: W1001 09:20:19.730625 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a7230d0_bbf6_4949_9af0_6489822f46f5.slice/crio-efd5973b6f48b4edc954cd478b922f10c3408075e0175b66f31bcccdf525fb97 WatchSource:0}: Error finding container efd5973b6f48b4edc954cd478b922f10c3408075e0175b66f31bcccdf525fb97: Status 404 returned error can't find the container with id efd5973b6f48b4edc954cd478b922f10c3408075e0175b66f31bcccdf525fb97 Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.864101 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" event={"ID":"7a7230d0-bbf6-4949-9af0-6489822f46f5","Type":"ContainerStarted","Data":"efd5973b6f48b4edc954cd478b922f10c3408075e0175b66f31bcccdf525fb97"} Oct 01 09:20:19 crc kubenswrapper[4983]: I1001 09:20:19.865307 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" event={"ID":"31e9843f-42b4-4229-bdfe-d64ccd1b37eb","Type":"ContainerStarted","Data":"11f798f567a3df0448939016a6e4bbaeab07a0601e7b0fb4b81dfbf3786e285f"} Oct 01 09:20:22 crc kubenswrapper[4983]: I1001 09:20:22.882561 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" event={"ID":"31e9843f-42b4-4229-bdfe-d64ccd1b37eb","Type":"ContainerStarted","Data":"bd4bbc9f6f94d5cd44e7b1475552d6da98276e6e26a6ddb0b7b2f411ce5ccb00"} Oct 01 09:20:22 crc kubenswrapper[4983]: I1001 09:20:22.882828 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:20:22 crc kubenswrapper[4983]: I1001 09:20:22.904769 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" podStartSLOduration=2.265644722 podStartE2EDuration="4.904740111s" podCreationTimestamp="2025-10-01 09:20:18 +0000 UTC" firstStartedPulling="2025-10-01 09:20:19.489925533 +0000 UTC m=+647.479154330" lastFinishedPulling="2025-10-01 09:20:22.129020912 +0000 UTC m=+650.118249719" observedRunningTime="2025-10-01 09:20:22.899447891 +0000 UTC m=+650.888676688" watchObservedRunningTime="2025-10-01 09:20:22.904740111 +0000 UTC m=+650.893968908" Oct 01 09:20:23 crc kubenswrapper[4983]: I1001 09:20:23.889217 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" event={"ID":"7a7230d0-bbf6-4949-9af0-6489822f46f5","Type":"ContainerStarted","Data":"a4dd647331f45e6490f0f428661dc3fd3d6cbaabacc5d696bb832af225d79ff3"} Oct 01 09:20:23 crc kubenswrapper[4983]: I1001 09:20:23.907938 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" podStartSLOduration=1.003996104 podStartE2EDuration="4.907914267s" podCreationTimestamp="2025-10-01 09:20:19 +0000 UTC" firstStartedPulling="2025-10-01 09:20:19.733161845 +0000 UTC m=+647.722390642" lastFinishedPulling="2025-10-01 09:20:23.637079998 +0000 UTC m=+651.626308805" observedRunningTime="2025-10-01 09:20:23.904520597 +0000 UTC m=+651.893749434" watchObservedRunningTime="2025-10-01 09:20:23.907914267 +0000 UTC m=+651.897143074" Oct 01 09:20:24 crc kubenswrapper[4983]: I1001 09:20:24.897845 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:39 crc kubenswrapper[4983]: I1001 09:20:39.492662 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-67c678544c-thtq2" Oct 01 09:20:59 crc kubenswrapper[4983]: I1001 09:20:59.268483 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7d59c47655-9kwhv" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.108284 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-q6g4h"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.110237 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.115114 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.115177 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.115318 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xmnx9" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.130896 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.131685 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.135791 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.149463 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228067 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-cert\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228136 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmdnm\" (UniqueName: \"kubernetes.io/projected/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-kube-api-access-qmdnm\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228161 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrcrs\" (UniqueName: \"kubernetes.io/projected/b5648467-cebb-4305-a03b-cfe484b92d9e-kube-api-access-mrcrs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228191 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-startup\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228210 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics-certs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228257 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-sockets\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228275 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-reloader\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228363 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228392 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-conf\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.228879 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-h7cfp"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.229663 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.231226 4983 reflector.go:561] object-"metallb-system"/"speaker-dockercfg-nfd9z": failed to list *v1.Secret: secrets "speaker-dockercfg-nfd9z" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:21:00 crc kubenswrapper[4983]: E1001 09:21:00.231256 4983 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"speaker-dockercfg-nfd9z\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"speaker-dockercfg-nfd9z\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.231341 4983 reflector.go:561] object-"metallb-system"/"speaker-certs-secret": failed to list *v1.Secret: secrets "speaker-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:21:00 crc kubenswrapper[4983]: E1001 09:21:00.231356 4983 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"speaker-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"speaker-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.231638 4983 reflector.go:561] object-"metallb-system"/"metallb-memberlist": failed to list *v1.Secret: secrets "metallb-memberlist" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:21:00 crc kubenswrapper[4983]: E1001 09:21:00.231745 4983 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-memberlist\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-memberlist\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.231831 4983 reflector.go:561] object-"metallb-system"/"metallb-excludel2": failed to list *v1.ConfigMap: configmaps "metallb-excludel2" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:21:00 crc kubenswrapper[4983]: E1001 09:21:00.231870 4983 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-excludel2\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"metallb-excludel2\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.261669 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-g55d4"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.262484 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.266356 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.271743 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-g55d4"] Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329710 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-startup\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329763 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics-certs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329785 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-sockets\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329800 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-reloader\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329849 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgflm\" (UniqueName: \"kubernetes.io/projected/13e9026f-afa7-4216-ae81-208fb18c8d08-kube-api-access-bgflm\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329900 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-metrics-certs\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329928 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329951 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.329975 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-conf\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330084 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-cert\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330185 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330635 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330643 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-conf\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330765 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-reloader\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330697 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-sockets\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.330208 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmdnm\" (UniqueName: \"kubernetes.io/projected/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-kube-api-access-qmdnm\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.331138 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrcrs\" (UniqueName: \"kubernetes.io/projected/b5648467-cebb-4305-a03b-cfe484b92d9e-kube-api-access-mrcrs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.331163 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b5648467-cebb-4305-a03b-cfe484b92d9e-frr-startup\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.335388 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-cert\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.339236 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5648467-cebb-4305-a03b-cfe484b92d9e-metrics-certs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.346365 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrcrs\" (UniqueName: \"kubernetes.io/projected/b5648467-cebb-4305-a03b-cfe484b92d9e-kube-api-access-mrcrs\") pod \"frr-k8s-q6g4h\" (UID: \"b5648467-cebb-4305-a03b-cfe484b92d9e\") " pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.349565 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmdnm\" (UniqueName: \"kubernetes.io/projected/97d2b4b2-bfd1-4007-949f-9f8fa01053a7-kube-api-access-qmdnm\") pod \"frr-k8s-webhook-server-5478bdb765-d7w8t\" (UID: \"97d2b4b2-bfd1-4007-949f-9f8fa01053a7\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.427568 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432059 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-metrics-certs\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432118 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgflm\" (UniqueName: \"kubernetes.io/projected/13e9026f-afa7-4216-ae81-208fb18c8d08-kube-api-access-bgflm\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432143 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smf7r\" (UniqueName: \"kubernetes.io/projected/b3825301-65ff-477d-b186-518fd24169a4-kube-api-access-smf7r\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432174 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-metrics-certs\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432407 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432512 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.432569 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-cert\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.444895 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.448013 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgflm\" (UniqueName: \"kubernetes.io/projected/13e9026f-afa7-4216-ae81-208fb18c8d08-kube-api-access-bgflm\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.534255 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-cert\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.534618 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-metrics-certs\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.534660 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smf7r\" (UniqueName: \"kubernetes.io/projected/b3825301-65ff-477d-b186-518fd24169a4-kube-api-access-smf7r\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.536386 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.539879 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-metrics-certs\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.549489 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3825301-65ff-477d-b186-518fd24169a4-cert\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.552070 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smf7r\" (UniqueName: \"kubernetes.io/projected/b3825301-65ff-477d-b186-518fd24169a4-kube-api-access-smf7r\") pod \"controller-5d688f5ffc-g55d4\" (UID: \"b3825301-65ff-477d-b186-518fd24169a4\") " pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.577142 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.631329 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t"] Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.641226 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97d2b4b2_bfd1_4007_949f_9f8fa01053a7.slice/crio-b2f282d983013558bb1c24cc46c2ef753abe4b0be88d4afe65ab2ecb174bf0e6 WatchSource:0}: Error finding container b2f282d983013558bb1c24cc46c2ef753abe4b0be88d4afe65ab2ecb174bf0e6: Status 404 returned error can't find the container with id b2f282d983013558bb1c24cc46c2ef753abe4b0be88d4afe65ab2ecb174bf0e6 Oct 01 09:21:00 crc kubenswrapper[4983]: I1001 09:21:00.745300 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-g55d4"] Oct 01 09:21:00 crc kubenswrapper[4983]: W1001 09:21:00.751622 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3825301_65ff_477d_b186_518fd24169a4.slice/crio-fdaa1642807c7bd5b7d336b58e9ccec9ef35add3ec6f10bb22d2095775416009 WatchSource:0}: Error finding container fdaa1642807c7bd5b7d336b58e9ccec9ef35add3ec6f10bb22d2095775416009: Status 404 returned error can't find the container with id fdaa1642807c7bd5b7d336b58e9ccec9ef35add3ec6f10bb22d2095775416009 Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.087707 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-g55d4" event={"ID":"b3825301-65ff-477d-b186-518fd24169a4","Type":"ContainerStarted","Data":"51cff5d22dfd9bdc62418d39b9a97459e8c7e4321fed8579b9f94c6a65409ba8"} Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.087757 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-g55d4" event={"ID":"b3825301-65ff-477d-b186-518fd24169a4","Type":"ContainerStarted","Data":"fdaa1642807c7bd5b7d336b58e9ccec9ef35add3ec6f10bb22d2095775416009"} Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.088900 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"6021899dc130c6a86a9738aa5e860f66f3843906ee3d102a96f37bf89da5e3d5"} Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.090027 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" event={"ID":"97d2b4b2-bfd1-4007-949f-9f8fa01053a7","Type":"ContainerStarted","Data":"b2f282d983013558bb1c24cc46c2ef753abe4b0be88d4afe65ab2ecb174bf0e6"} Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.425927 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 01 09:21:01 crc kubenswrapper[4983]: E1001 09:21:01.433180 4983 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: failed to sync secret cache: timed out waiting for the condition Oct 01 09:21:01 crc kubenswrapper[4983]: E1001 09:21:01.433228 4983 configmap.go:193] Couldn't get configMap metallb-system/metallb-excludel2: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:21:01 crc kubenswrapper[4983]: E1001 09:21:01.433257 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist podName:13e9026f-afa7-4216-ae81-208fb18c8d08 nodeName:}" failed. No retries permitted until 2025-10-01 09:21:01.933238229 +0000 UTC m=+689.922467026 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist") pod "speaker-h7cfp" (UID: "13e9026f-afa7-4216-ae81-208fb18c8d08") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:21:01 crc kubenswrapper[4983]: E1001 09:21:01.433293 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2 podName:13e9026f-afa7-4216-ae81-208fb18c8d08 nodeName:}" failed. No retries permitted until 2025-10-01 09:21:01.93327643 +0000 UTC m=+689.922505227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metallb-excludel2" (UniqueName: "kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2") pod "speaker-h7cfp" (UID: "13e9026f-afa7-4216-ae81-208fb18c8d08") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.438552 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-metrics-certs\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.529411 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.563405 4983 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-nfd9z" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.778548 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.953748 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.953873 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.954527 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/13e9026f-afa7-4216-ae81-208fb18c8d08-metallb-excludel2\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:01 crc kubenswrapper[4983]: I1001 09:21:01.956473 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/13e9026f-afa7-4216-ae81-208fb18c8d08-memberlist\") pod \"speaker-h7cfp\" (UID: \"13e9026f-afa7-4216-ae81-208fb18c8d08\") " pod="metallb-system/speaker-h7cfp" Oct 01 09:21:02 crc kubenswrapper[4983]: I1001 09:21:02.041044 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h7cfp" Oct 01 09:21:02 crc kubenswrapper[4983]: W1001 09:21:02.073536 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13e9026f_afa7_4216_ae81_208fb18c8d08.slice/crio-5aec467f31055eb5510849a39424d451eb957c99d3a033b616620c13792f49a5 WatchSource:0}: Error finding container 5aec467f31055eb5510849a39424d451eb957c99d3a033b616620c13792f49a5: Status 404 returned error can't find the container with id 5aec467f31055eb5510849a39424d451eb957c99d3a033b616620c13792f49a5 Oct 01 09:21:02 crc kubenswrapper[4983]: I1001 09:21:02.098373 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7cfp" event={"ID":"13e9026f-afa7-4216-ae81-208fb18c8d08","Type":"ContainerStarted","Data":"5aec467f31055eb5510849a39424d451eb957c99d3a033b616620c13792f49a5"} Oct 01 09:21:03 crc kubenswrapper[4983]: I1001 09:21:03.107081 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7cfp" event={"ID":"13e9026f-afa7-4216-ae81-208fb18c8d08","Type":"ContainerStarted","Data":"409e5df75f9ea36a14b10379bcdada9d2efc0c55e19869bd667a92810dd79079"} Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.123223 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-g55d4" event={"ID":"b3825301-65ff-477d-b186-518fd24169a4","Type":"ContainerStarted","Data":"13159ec323ff0798f12e3ddcf861e08b9858c32109be51a0d81401d978148517"} Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.123575 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.126339 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h7cfp" event={"ID":"13e9026f-afa7-4216-ae81-208fb18c8d08","Type":"ContainerStarted","Data":"485423c0dce14b4bf2ff546fa4d98daf0e000e77cc7f6df14a745be1ea4bc7d1"} Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.126578 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-h7cfp" Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.142321 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-g55d4" podStartSLOduration=1.5252369749999999 podStartE2EDuration="5.142300071s" podCreationTimestamp="2025-10-01 09:21:00 +0000 UTC" firstStartedPulling="2025-10-01 09:21:00.907339291 +0000 UTC m=+688.896568108" lastFinishedPulling="2025-10-01 09:21:04.524402407 +0000 UTC m=+692.513631204" observedRunningTime="2025-10-01 09:21:05.139955849 +0000 UTC m=+693.129184656" watchObservedRunningTime="2025-10-01 09:21:05.142300071 +0000 UTC m=+693.131528868" Oct 01 09:21:05 crc kubenswrapper[4983]: I1001 09:21:05.171563 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-h7cfp" podStartSLOduration=3.175356166 podStartE2EDuration="5.171543798s" podCreationTimestamp="2025-10-01 09:21:00 +0000 UTC" firstStartedPulling="2025-10-01 09:21:02.513991287 +0000 UTC m=+690.503220084" lastFinishedPulling="2025-10-01 09:21:04.510178919 +0000 UTC m=+692.499407716" observedRunningTime="2025-10-01 09:21:05.168782155 +0000 UTC m=+693.158010962" watchObservedRunningTime="2025-10-01 09:21:05.171543798 +0000 UTC m=+693.160772595" Oct 01 09:21:08 crc kubenswrapper[4983]: I1001 09:21:08.147772 4983 generic.go:334] "Generic (PLEG): container finished" podID="b5648467-cebb-4305-a03b-cfe484b92d9e" containerID="68a714fc854896f9eb2efa92c76dd9cbb3001e94bf5f843d8f6b16cbe1a7e8dc" exitCode=0 Oct 01 09:21:08 crc kubenswrapper[4983]: I1001 09:21:08.147887 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerDied","Data":"68a714fc854896f9eb2efa92c76dd9cbb3001e94bf5f843d8f6b16cbe1a7e8dc"} Oct 01 09:21:08 crc kubenswrapper[4983]: I1001 09:21:08.150010 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" event={"ID":"97d2b4b2-bfd1-4007-949f-9f8fa01053a7","Type":"ContainerStarted","Data":"4b61c11c3158b30b03af65e79a76c4c44f8e158a5dc43a029f906c9ba48dd4ce"} Oct 01 09:21:08 crc kubenswrapper[4983]: I1001 09:21:08.150153 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:08 crc kubenswrapper[4983]: I1001 09:21:08.206289 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" podStartSLOduration=1.613509053 podStartE2EDuration="8.206265406s" podCreationTimestamp="2025-10-01 09:21:00 +0000 UTC" firstStartedPulling="2025-10-01 09:21:00.644070553 +0000 UTC m=+688.633299350" lastFinishedPulling="2025-10-01 09:21:07.236826906 +0000 UTC m=+695.226055703" observedRunningTime="2025-10-01 09:21:08.204496939 +0000 UTC m=+696.193725736" watchObservedRunningTime="2025-10-01 09:21:08.206265406 +0000 UTC m=+696.195494223" Oct 01 09:21:09 crc kubenswrapper[4983]: I1001 09:21:09.159228 4983 generic.go:334] "Generic (PLEG): container finished" podID="b5648467-cebb-4305-a03b-cfe484b92d9e" containerID="0bef36cfa6b053d81aaab637b4007a7192ee0ccb1f69b4a9c10a48e012600dc1" exitCode=0 Oct 01 09:21:09 crc kubenswrapper[4983]: I1001 09:21:09.159361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerDied","Data":"0bef36cfa6b053d81aaab637b4007a7192ee0ccb1f69b4a9c10a48e012600dc1"} Oct 01 09:21:10 crc kubenswrapper[4983]: I1001 09:21:10.165627 4983 generic.go:334] "Generic (PLEG): container finished" podID="b5648467-cebb-4305-a03b-cfe484b92d9e" containerID="31100f843fbbfa679584fb6a8ee9cb495c081962f682b9fa05475144b6601238" exitCode=0 Oct 01 09:21:10 crc kubenswrapper[4983]: I1001 09:21:10.165673 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerDied","Data":"31100f843fbbfa679584fb6a8ee9cb495c081962f682b9fa05475144b6601238"} Oct 01 09:21:10 crc kubenswrapper[4983]: I1001 09:21:10.581190 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-g55d4" Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.178189 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"3b3c12661e7500d8a887e66e3439db061f10c52edad14435cbd7e2976bd81638"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179140 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179215 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"42d07b5c4e052d51260639b4b7d8a45a0f93143974251783c6dd0baae900c608"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179280 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"365435d0c20a9b34e28c5217f312c86a9116c380925e79b7fc11901574e5a268"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179347 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"9b3f9fba7573bc233a595e426ba77c7f26e1c715bb4e8db050f36e8ff07e79d5"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179410 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"e9279332b01e998490b2e47f5194a12783adb7a198edfd194e475b2e24ca0c39"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.179468 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-q6g4h" event={"ID":"b5648467-cebb-4305-a03b-cfe484b92d9e","Type":"ContainerStarted","Data":"d94a51ed3f494d5736388b59307ab88f85e5f7d9d7b2950a44007c668949f0f4"} Oct 01 09:21:11 crc kubenswrapper[4983]: I1001 09:21:11.201155 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-q6g4h" podStartSLOduration=4.545246931 podStartE2EDuration="11.201134163s" podCreationTimestamp="2025-10-01 09:21:00 +0000 UTC" firstStartedPulling="2025-10-01 09:21:00.595623345 +0000 UTC m=+688.584852142" lastFinishedPulling="2025-10-01 09:21:07.251510577 +0000 UTC m=+695.240739374" observedRunningTime="2025-10-01 09:21:11.19874468 +0000 UTC m=+699.187973507" watchObservedRunningTime="2025-10-01 09:21:11.201134163 +0000 UTC m=+699.190362960" Oct 01 09:21:12 crc kubenswrapper[4983]: I1001 09:21:12.045220 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-h7cfp" Oct 01 09:21:15 crc kubenswrapper[4983]: I1001 09:21:15.428823 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:15 crc kubenswrapper[4983]: I1001 09:21:15.474220 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:20 crc kubenswrapper[4983]: I1001 09:21:20.431279 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-q6g4h" Oct 01 09:21:20 crc kubenswrapper[4983]: I1001 09:21:20.454153 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-d7w8t" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.557772 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-2m9j5"] Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.558912 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.560900 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.561034 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.561152 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-tghvq" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.564069 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-2m9j5"] Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.704174 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl577\" (UniqueName: \"kubernetes.io/projected/a695e1aa-e7cd-40fd-bacb-912fc3881f9c-kube-api-access-wl577\") pod \"mariadb-operator-index-2m9j5\" (UID: \"a695e1aa-e7cd-40fd-bacb-912fc3881f9c\") " pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.805303 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl577\" (UniqueName: \"kubernetes.io/projected/a695e1aa-e7cd-40fd-bacb-912fc3881f9c-kube-api-access-wl577\") pod \"mariadb-operator-index-2m9j5\" (UID: \"a695e1aa-e7cd-40fd-bacb-912fc3881f9c\") " pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.826513 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl577\" (UniqueName: \"kubernetes.io/projected/a695e1aa-e7cd-40fd-bacb-912fc3881f9c-kube-api-access-wl577\") pod \"mariadb-operator-index-2m9j5\" (UID: \"a695e1aa-e7cd-40fd-bacb-912fc3881f9c\") " pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:21 crc kubenswrapper[4983]: I1001 09:21:21.879513 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:22 crc kubenswrapper[4983]: I1001 09:21:22.057584 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-2m9j5"] Oct 01 09:21:22 crc kubenswrapper[4983]: W1001 09:21:22.064632 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda695e1aa_e7cd_40fd_bacb_912fc3881f9c.slice/crio-2a442adbbc8e60e33525723bd23954aac93645243249761bda215cc97b12b114 WatchSource:0}: Error finding container 2a442adbbc8e60e33525723bd23954aac93645243249761bda215cc97b12b114: Status 404 returned error can't find the container with id 2a442adbbc8e60e33525723bd23954aac93645243249761bda215cc97b12b114 Oct 01 09:21:22 crc kubenswrapper[4983]: I1001 09:21:22.247332 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-2m9j5" event={"ID":"a695e1aa-e7cd-40fd-bacb-912fc3881f9c","Type":"ContainerStarted","Data":"2a442adbbc8e60e33525723bd23954aac93645243249761bda215cc97b12b114"} Oct 01 09:21:24 crc kubenswrapper[4983]: I1001 09:21:24.263522 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-2m9j5" event={"ID":"a695e1aa-e7cd-40fd-bacb-912fc3881f9c","Type":"ContainerStarted","Data":"5924a7ef9715fefafe72ee63f4855648bdf668694c9b92a2d018917499820c20"} Oct 01 09:21:24 crc kubenswrapper[4983]: I1001 09:21:24.280773 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-2m9j5" podStartSLOduration=2.080768479 podStartE2EDuration="3.280752966s" podCreationTimestamp="2025-10-01 09:21:21 +0000 UTC" firstStartedPulling="2025-10-01 09:21:22.066732825 +0000 UTC m=+710.055961622" lastFinishedPulling="2025-10-01 09:21:23.266717302 +0000 UTC m=+711.255946109" observedRunningTime="2025-10-01 09:21:24.280598872 +0000 UTC m=+712.269827699" watchObservedRunningTime="2025-10-01 09:21:24.280752966 +0000 UTC m=+712.269981763" Oct 01 09:21:31 crc kubenswrapper[4983]: I1001 09:21:31.879857 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:31 crc kubenswrapper[4983]: I1001 09:21:31.880360 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:31 crc kubenswrapper[4983]: I1001 09:21:31.917591 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:32 crc kubenswrapper[4983]: I1001 09:21:32.326936 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-2m9j5" Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.820527 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2"] Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.822282 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.824203 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.830635 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2"] Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.967113 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.967162 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rk2w\" (UniqueName: \"kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:34 crc kubenswrapper[4983]: I1001 09:21:34.967204 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.069041 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rk2w\" (UniqueName: \"kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.069189 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.069393 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.069767 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.069950 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.088419 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rk2w\" (UniqueName: \"kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w\") pod \"27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.149168 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:35 crc kubenswrapper[4983]: I1001 09:21:35.551274 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2"] Oct 01 09:21:35 crc kubenswrapper[4983]: W1001 09:21:35.560293 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a8ec064_7744_4a7c_8479_04b47b73712a.slice/crio-cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6 WatchSource:0}: Error finding container cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6: Status 404 returned error can't find the container with id cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6 Oct 01 09:21:36 crc kubenswrapper[4983]: I1001 09:21:36.329874 4983 generic.go:334] "Generic (PLEG): container finished" podID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerID="62b078277a4a52399cb040b58f268bf6adb125f98e5aa26dfb0b06ac83fe1609" exitCode=0 Oct 01 09:21:36 crc kubenswrapper[4983]: I1001 09:21:36.329941 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" event={"ID":"7a8ec064-7744-4a7c-8479-04b47b73712a","Type":"ContainerDied","Data":"62b078277a4a52399cb040b58f268bf6adb125f98e5aa26dfb0b06ac83fe1609"} Oct 01 09:21:36 crc kubenswrapper[4983]: I1001 09:21:36.330143 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" event={"ID":"7a8ec064-7744-4a7c-8479-04b47b73712a","Type":"ContainerStarted","Data":"cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6"} Oct 01 09:21:37 crc kubenswrapper[4983]: I1001 09:21:37.337745 4983 generic.go:334] "Generic (PLEG): container finished" podID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerID="3e99c7a883c1d389ba859d0ea8f0856a6ab608e8b8ddee483263f3cedbb5ba67" exitCode=0 Oct 01 09:21:37 crc kubenswrapper[4983]: I1001 09:21:37.337858 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" event={"ID":"7a8ec064-7744-4a7c-8479-04b47b73712a","Type":"ContainerDied","Data":"3e99c7a883c1d389ba859d0ea8f0856a6ab608e8b8ddee483263f3cedbb5ba67"} Oct 01 09:21:38 crc kubenswrapper[4983]: I1001 09:21:38.344332 4983 generic.go:334] "Generic (PLEG): container finished" podID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerID="261f072e917f2907713ca826d80d918520a4f0c56f587b64dbcf9aee956ba464" exitCode=0 Oct 01 09:21:38 crc kubenswrapper[4983]: I1001 09:21:38.344383 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" event={"ID":"7a8ec064-7744-4a7c-8479-04b47b73712a","Type":"ContainerDied","Data":"261f072e917f2907713ca826d80d918520a4f0c56f587b64dbcf9aee956ba464"} Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.589297 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.726087 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rk2w\" (UniqueName: \"kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w\") pod \"7a8ec064-7744-4a7c-8479-04b47b73712a\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.726248 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle\") pod \"7a8ec064-7744-4a7c-8479-04b47b73712a\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.726292 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util\") pod \"7a8ec064-7744-4a7c-8479-04b47b73712a\" (UID: \"7a8ec064-7744-4a7c-8479-04b47b73712a\") " Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.727099 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle" (OuterVolumeSpecName: "bundle") pod "7a8ec064-7744-4a7c-8479-04b47b73712a" (UID: "7a8ec064-7744-4a7c-8479-04b47b73712a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.732651 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w" (OuterVolumeSpecName: "kube-api-access-7rk2w") pod "7a8ec064-7744-4a7c-8479-04b47b73712a" (UID: "7a8ec064-7744-4a7c-8479-04b47b73712a"). InnerVolumeSpecName "kube-api-access-7rk2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.741319 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util" (OuterVolumeSpecName: "util") pod "7a8ec064-7744-4a7c-8479-04b47b73712a" (UID: "7a8ec064-7744-4a7c-8479-04b47b73712a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.827326 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.827351 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a8ec064-7744-4a7c-8479-04b47b73712a-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4983]: I1001 09:21:39.827360 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rk2w\" (UniqueName: \"kubernetes.io/projected/7a8ec064-7744-4a7c-8479-04b47b73712a-kube-api-access-7rk2w\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:40 crc kubenswrapper[4983]: I1001 09:21:40.356325 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" event={"ID":"7a8ec064-7744-4a7c-8479-04b47b73712a","Type":"ContainerDied","Data":"cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6"} Oct 01 09:21:40 crc kubenswrapper[4983]: I1001 09:21:40.356364 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2" Oct 01 09:21:40 crc kubenswrapper[4983]: I1001 09:21:40.356375 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb7a78e33a3629605978c68710424e2d67779823327c518446aa533fb33d28b6" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.384590 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh"] Oct 01 09:21:45 crc kubenswrapper[4983]: E1001 09:21:45.385378 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="pull" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.385394 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="pull" Oct 01 09:21:45 crc kubenswrapper[4983]: E1001 09:21:45.385419 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="extract" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.385427 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="extract" Oct 01 09:21:45 crc kubenswrapper[4983]: E1001 09:21:45.385439 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="util" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.385447 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="util" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.385569 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8ec064-7744-4a7c-8479-04b47b73712a" containerName="extract" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.386333 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.387756 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.390894 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4275j" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.391550 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.396281 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh"] Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.501170 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.501238 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-webhook-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.501279 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znhpr\" (UniqueName: \"kubernetes.io/projected/922be10f-7fd4-413d-a422-7b004b1be5a5-kube-api-access-znhpr\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.602598 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.602674 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-webhook-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.602716 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znhpr\" (UniqueName: \"kubernetes.io/projected/922be10f-7fd4-413d-a422-7b004b1be5a5-kube-api-access-znhpr\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.610102 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-webhook-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.610144 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/922be10f-7fd4-413d-a422-7b004b1be5a5-apiservice-cert\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.617330 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znhpr\" (UniqueName: \"kubernetes.io/projected/922be10f-7fd4-413d-a422-7b004b1be5a5-kube-api-access-znhpr\") pod \"mariadb-operator-controller-manager-79c894898-kq4lh\" (UID: \"922be10f-7fd4-413d-a422-7b004b1be5a5\") " pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.711460 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:45 crc kubenswrapper[4983]: I1001 09:21:45.908954 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh"] Oct 01 09:21:46 crc kubenswrapper[4983]: I1001 09:21:46.392879 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" event={"ID":"922be10f-7fd4-413d-a422-7b004b1be5a5","Type":"ContainerStarted","Data":"af97c7b828163086329c2b4a032b4b8d004d7198a7b3edf1b5dbc22fd750d7af"} Oct 01 09:21:49 crc kubenswrapper[4983]: I1001 09:21:49.410955 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" event={"ID":"922be10f-7fd4-413d-a422-7b004b1be5a5","Type":"ContainerStarted","Data":"7eb56c66f2ecedae5ce3a5ba13e732ffbff2e22f0c08380fc121c157c97acb62"} Oct 01 09:21:51 crc kubenswrapper[4983]: I1001 09:21:51.422281 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" event={"ID":"922be10f-7fd4-413d-a422-7b004b1be5a5","Type":"ContainerStarted","Data":"827e57ad537c4bb5737febbb22b4a946fb3cfc78d335070a0421bf552252b933"} Oct 01 09:21:51 crc kubenswrapper[4983]: I1001 09:21:51.422633 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:51 crc kubenswrapper[4983]: I1001 09:21:51.443502 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" podStartSLOduration=1.202763094 podStartE2EDuration="6.443485655s" podCreationTimestamp="2025-10-01 09:21:45 +0000 UTC" firstStartedPulling="2025-10-01 09:21:45.929478882 +0000 UTC m=+733.918707669" lastFinishedPulling="2025-10-01 09:21:51.170201433 +0000 UTC m=+739.159430230" observedRunningTime="2025-10-01 09:21:51.442489573 +0000 UTC m=+739.431718390" watchObservedRunningTime="2025-10-01 09:21:51.443485655 +0000 UTC m=+739.432714452" Oct 01 09:21:55 crc kubenswrapper[4983]: I1001 09:21:55.716877 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c894898-kq4lh" Oct 01 09:21:57 crc kubenswrapper[4983]: I1001 09:21:57.868874 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:21:57 crc kubenswrapper[4983]: I1001 09:21:57.870095 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:21:57 crc kubenswrapper[4983]: I1001 09:21:57.874293 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-drqjn" Oct 01 09:21:57 crc kubenswrapper[4983]: I1001 09:21:57.905530 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:21:57 crc kubenswrapper[4983]: I1001 09:21:57.949280 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh2qt\" (UniqueName: \"kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt\") pod \"infra-operator-index-8w9dd\" (UID: \"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab\") " pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:21:58 crc kubenswrapper[4983]: I1001 09:21:58.050600 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh2qt\" (UniqueName: \"kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt\") pod \"infra-operator-index-8w9dd\" (UID: \"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab\") " pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:21:58 crc kubenswrapper[4983]: I1001 09:21:58.099335 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh2qt\" (UniqueName: \"kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt\") pod \"infra-operator-index-8w9dd\" (UID: \"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab\") " pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:21:58 crc kubenswrapper[4983]: I1001 09:21:58.187771 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:21:58 crc kubenswrapper[4983]: I1001 09:21:58.575693 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.465394 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8w9dd" event={"ID":"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab","Type":"ContainerStarted","Data":"3b4183c71e4a59959a56a12584ad46d12c23ffc380617955c5e6a90fd59cb9bb"} Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.535320 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.535587 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" podUID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" containerName="controller-manager" containerID="cri-o://7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e" gracePeriod=30 Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.547247 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.547448 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" containerName="route-controller-manager" containerID="cri-o://d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee" gracePeriod=30 Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.931621 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:21:59 crc kubenswrapper[4983]: I1001 09:21:59.937186 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076290 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mmdb\" (UniqueName: \"kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb\") pod \"d081f5a5-7101-403d-82f5-4395d1db959e\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076386 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config\") pod \"d081f5a5-7101-403d-82f5-4395d1db959e\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076418 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca\") pod \"d081f5a5-7101-403d-82f5-4395d1db959e\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076441 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca\") pod \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076454 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config\") pod \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076479 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert\") pod \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076497 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles\") pod \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076518 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert\") pod \"d081f5a5-7101-403d-82f5-4395d1db959e\" (UID: \"d081f5a5-7101-403d-82f5-4395d1db959e\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.076559 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-262xt\" (UniqueName: \"kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt\") pod \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\" (UID: \"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a\") " Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.077290 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" (UID: "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.077389 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca" (OuterVolumeSpecName: "client-ca") pod "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" (UID: "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.077677 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config" (OuterVolumeSpecName: "config") pod "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" (UID: "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.077682 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca" (OuterVolumeSpecName: "client-ca") pod "d081f5a5-7101-403d-82f5-4395d1db959e" (UID: "d081f5a5-7101-403d-82f5-4395d1db959e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.077865 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config" (OuterVolumeSpecName: "config") pod "d081f5a5-7101-403d-82f5-4395d1db959e" (UID: "d081f5a5-7101-403d-82f5-4395d1db959e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.081220 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" (UID: "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.081518 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d081f5a5-7101-403d-82f5-4395d1db959e" (UID: "d081f5a5-7101-403d-82f5-4395d1db959e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.081518 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt" (OuterVolumeSpecName: "kube-api-access-262xt") pod "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" (UID: "f1f1eb5b-6daf-49db-a36d-32efc8e8f09a"). InnerVolumeSpecName "kube-api-access-262xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.081581 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb" (OuterVolumeSpecName: "kube-api-access-9mmdb") pod "d081f5a5-7101-403d-82f5-4395d1db959e" (UID: "d081f5a5-7101-403d-82f5-4395d1db959e"). InnerVolumeSpecName "kube-api-access-9mmdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177824 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d081f5a5-7101-403d-82f5-4395d1db959e-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177867 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-262xt\" (UniqueName: \"kubernetes.io/projected/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-kube-api-access-262xt\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177879 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mmdb\" (UniqueName: \"kubernetes.io/projected/d081f5a5-7101-403d-82f5-4395d1db959e-kube-api-access-9mmdb\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177891 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177901 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d081f5a5-7101-403d-82f5-4395d1db959e-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177915 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177924 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177934 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.177943 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.472057 4983 generic.go:334] "Generic (PLEG): container finished" podID="d081f5a5-7101-403d-82f5-4395d1db959e" containerID="d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee" exitCode=0 Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.472378 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" event={"ID":"d081f5a5-7101-403d-82f5-4395d1db959e","Type":"ContainerDied","Data":"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee"} Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.472410 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" event={"ID":"d081f5a5-7101-403d-82f5-4395d1db959e","Type":"ContainerDied","Data":"d5eda2896009250a7ad4c1a1b6ab4d275f609ccdce6ac7aecc6b2c7e31914ff2"} Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.472430 4983 scope.go:117] "RemoveContainer" containerID="d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.472504 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.475077 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8w9dd" event={"ID":"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab","Type":"ContainerStarted","Data":"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233"} Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.476955 4983 generic.go:334] "Generic (PLEG): container finished" podID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" containerID="7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e" exitCode=0 Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.476992 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" event={"ID":"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a","Type":"ContainerDied","Data":"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e"} Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.477038 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" event={"ID":"f1f1eb5b-6daf-49db-a36d-32efc8e8f09a","Type":"ContainerDied","Data":"4d7d0f2107916625ba2e4d91850ae896377dd8a7f4d1f5284cf1ff8930f338d2"} Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.477025 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-wnxqd" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.490595 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-8w9dd" podStartSLOduration=2.497017293 podStartE2EDuration="3.490575391s" podCreationTimestamp="2025-10-01 09:21:57 +0000 UTC" firstStartedPulling="2025-10-01 09:21:58.582594412 +0000 UTC m=+746.571823209" lastFinishedPulling="2025-10-01 09:21:59.57615251 +0000 UTC m=+747.565381307" observedRunningTime="2025-10-01 09:22:00.488372399 +0000 UTC m=+748.477601216" watchObservedRunningTime="2025-10-01 09:22:00.490575391 +0000 UTC m=+748.479804198" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.493069 4983 scope.go:117] "RemoveContainer" containerID="d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee" Oct 01 09:22:00 crc kubenswrapper[4983]: E1001 09:22:00.493788 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee\": container with ID starting with d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee not found: ID does not exist" containerID="d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.493853 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee"} err="failed to get container status \"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee\": rpc error: code = NotFound desc = could not find container \"d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee\": container with ID starting with d913886fb905246dcd01b011c47b59cfc8f284b6ecd9886d3ed2ae5a2d5317ee not found: ID does not exist" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.493902 4983 scope.go:117] "RemoveContainer" containerID="7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.507720 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.511968 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ptgcj"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.512559 4983 scope.go:117] "RemoveContainer" containerID="7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e" Oct 01 09:22:00 crc kubenswrapper[4983]: E1001 09:22:00.513014 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e\": container with ID starting with 7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e not found: ID does not exist" containerID="7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.513045 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e"} err="failed to get container status \"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e\": rpc error: code = NotFound desc = could not find container \"7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e\": container with ID starting with 7736560bf200111f354059fa9410266c47a0b65424e713572de4947145f6437e not found: ID does not exist" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.515605 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.518680 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-wnxqd"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.720738 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" path="/var/lib/kubelet/pods/d081f5a5-7101-403d-82f5-4395d1db959e/volumes" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.721549 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" path="/var/lib/kubelet/pods/f1f1eb5b-6daf-49db-a36d-32efc8e8f09a/volumes" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.844637 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.862260 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-949548d95-2f48w"] Oct 01 09:22:00 crc kubenswrapper[4983]: E1001 09:22:00.862539 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" containerName="route-controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.862559 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" containerName="route-controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: E1001 09:22:00.862577 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" containerName="controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.862583 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" containerName="controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.862680 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f1eb5b-6daf-49db-a36d-32efc8e8f09a" containerName="controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.862693 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d081f5a5-7101-403d-82f5-4395d1db959e" containerName="route-controller-manager" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.863091 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.865296 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.866560 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.866561 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.868304 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.868656 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.869520 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.869955 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.872605 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.880922 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.881060 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.881283 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.881410 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.881442 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.881579 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.883458 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.896894 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.916301 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-949548d95-2f48w"] Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987017 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfzbc\" (UniqueName: \"kubernetes.io/projected/867184e8-80bf-4ebb-8da7-46c30129b05d-kube-api-access-gfzbc\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987125 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-config\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987158 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/867184e8-80bf-4ebb-8da7-46c30129b05d-serving-cert\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987176 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm52g\" (UniqueName: \"kubernetes.io/projected/8854ef5f-55db-4260-8b55-74460585508f-kube-api-access-gm52g\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987207 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-proxy-ca-bundles\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987227 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-client-ca\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987309 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-config\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987325 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-client-ca\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:00 crc kubenswrapper[4983]: I1001 09:22:00.987369 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8854ef5f-55db-4260-8b55-74460585508f-serving-cert\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088015 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8854ef5f-55db-4260-8b55-74460585508f-serving-cert\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088078 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfzbc\" (UniqueName: \"kubernetes.io/projected/867184e8-80bf-4ebb-8da7-46c30129b05d-kube-api-access-gfzbc\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088122 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-config\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088185 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/867184e8-80bf-4ebb-8da7-46c30129b05d-serving-cert\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088299 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm52g\" (UniqueName: \"kubernetes.io/projected/8854ef5f-55db-4260-8b55-74460585508f-kube-api-access-gm52g\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088341 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-proxy-ca-bundles\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088716 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-client-ca\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088782 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-config\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.088816 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-client-ca\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.089619 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-client-ca\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.090053 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-proxy-ca-bundles\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.090254 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-config\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.090371 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8854ef5f-55db-4260-8b55-74460585508f-client-ca\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.090379 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/867184e8-80bf-4ebb-8da7-46c30129b05d-config\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.091473 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8854ef5f-55db-4260-8b55-74460585508f-serving-cert\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.092008 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/867184e8-80bf-4ebb-8da7-46c30129b05d-serving-cert\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.105918 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm52g\" (UniqueName: \"kubernetes.io/projected/8854ef5f-55db-4260-8b55-74460585508f-kube-api-access-gm52g\") pod \"route-controller-manager-7b58b4d849-9nzrw\" (UID: \"8854ef5f-55db-4260-8b55-74460585508f\") " pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.108621 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfzbc\" (UniqueName: \"kubernetes.io/projected/867184e8-80bf-4ebb-8da7-46c30129b05d-kube-api-access-gfzbc\") pod \"controller-manager-949548d95-2f48w\" (UID: \"867184e8-80bf-4ebb-8da7-46c30129b05d\") " pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.182053 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.196001 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.452213 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-fxrbl"] Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.453297 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.458837 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-fxrbl"] Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.568258 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw"] Oct 01 09:22:01 crc kubenswrapper[4983]: W1001 09:22:01.573258 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8854ef5f_55db_4260_8b55_74460585508f.slice/crio-cd5a2606c06e66769fb765fe2a2fea49d67d00ee2b16524b48a3a25afacfbe7c WatchSource:0}: Error finding container cd5a2606c06e66769fb765fe2a2fea49d67d00ee2b16524b48a3a25afacfbe7c: Status 404 returned error can't find the container with id cd5a2606c06e66769fb765fe2a2fea49d67d00ee2b16524b48a3a25afacfbe7c Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.594701 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb5np\" (UniqueName: \"kubernetes.io/projected/a927379f-96be-4580-ad23-d807db431881-kube-api-access-xb5np\") pod \"infra-operator-index-fxrbl\" (UID: \"a927379f-96be-4580-ad23-d807db431881\") " pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.608746 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-949548d95-2f48w"] Oct 01 09:22:01 crc kubenswrapper[4983]: W1001 09:22:01.614819 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod867184e8_80bf_4ebb_8da7_46c30129b05d.slice/crio-59586c74dfd53462f54c5f36b5ca071550f11567353e551d6cd5d0f8c2eb5979 WatchSource:0}: Error finding container 59586c74dfd53462f54c5f36b5ca071550f11567353e551d6cd5d0f8c2eb5979: Status 404 returned error can't find the container with id 59586c74dfd53462f54c5f36b5ca071550f11567353e551d6cd5d0f8c2eb5979 Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.695751 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb5np\" (UniqueName: \"kubernetes.io/projected/a927379f-96be-4580-ad23-d807db431881-kube-api-access-xb5np\") pod \"infra-operator-index-fxrbl\" (UID: \"a927379f-96be-4580-ad23-d807db431881\") " pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.714483 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb5np\" (UniqueName: \"kubernetes.io/projected/a927379f-96be-4580-ad23-d807db431881-kube-api-access-xb5np\") pod \"infra-operator-index-fxrbl\" (UID: \"a927379f-96be-4580-ad23-d807db431881\") " pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.774035 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:01 crc kubenswrapper[4983]: I1001 09:22:01.992698 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-fxrbl"] Oct 01 09:22:01 crc kubenswrapper[4983]: W1001 09:22:01.998524 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda927379f_96be_4580_ad23_d807db431881.slice/crio-01c7f005941657c2d2eedda7aa9d63a634c1e97a4e0215e953f85b1711322e71 WatchSource:0}: Error finding container 01c7f005941657c2d2eedda7aa9d63a634c1e97a4e0215e953f85b1711322e71: Status 404 returned error can't find the container with id 01c7f005941657c2d2eedda7aa9d63a634c1e97a4e0215e953f85b1711322e71 Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.491240 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-fxrbl" event={"ID":"a927379f-96be-4580-ad23-d807db431881","Type":"ContainerStarted","Data":"01c7f005941657c2d2eedda7aa9d63a634c1e97a4e0215e953f85b1711322e71"} Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.492767 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" event={"ID":"8854ef5f-55db-4260-8b55-74460585508f","Type":"ContainerStarted","Data":"1e96da584400f08aa14168e4376e40643f65fe04e9dd83bca4094e4f08aa90e9"} Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.492795 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" event={"ID":"8854ef5f-55db-4260-8b55-74460585508f","Type":"ContainerStarted","Data":"cd5a2606c06e66769fb765fe2a2fea49d67d00ee2b16524b48a3a25afacfbe7c"} Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.492856 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.494614 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" event={"ID":"867184e8-80bf-4ebb-8da7-46c30129b05d","Type":"ContainerStarted","Data":"157d87d6b22b5b56c5299902d646960d92d4b1dc026f52078d3924d035eacaa2"} Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.494644 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" event={"ID":"867184e8-80bf-4ebb-8da7-46c30129b05d","Type":"ContainerStarted","Data":"59586c74dfd53462f54c5f36b5ca071550f11567353e551d6cd5d0f8c2eb5979"} Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.494930 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.496215 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-8w9dd" podUID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" containerName="registry-server" containerID="cri-o://fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233" gracePeriod=2 Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.499667 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.500710 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.513532 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7b58b4d849-9nzrw" podStartSLOduration=3.513516927 podStartE2EDuration="3.513516927s" podCreationTimestamp="2025-10-01 09:21:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:22:02.511612412 +0000 UTC m=+750.500841209" watchObservedRunningTime="2025-10-01 09:22:02.513516927 +0000 UTC m=+750.502745714" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.559024 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-949548d95-2f48w" podStartSLOduration=3.559004732 podStartE2EDuration="3.559004732s" podCreationTimestamp="2025-10-01 09:21:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:22:02.538455976 +0000 UTC m=+750.527684783" watchObservedRunningTime="2025-10-01 09:22:02.559004732 +0000 UTC m=+750.548233529" Oct 01 09:22:02 crc kubenswrapper[4983]: I1001 09:22:02.894208 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.013344 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh2qt\" (UniqueName: \"kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt\") pod \"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab\" (UID: \"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab\") " Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.018899 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt" (OuterVolumeSpecName: "kube-api-access-lh2qt") pod "e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" (UID: "e0f87abe-ea86-4f2b-be2f-4211bd4c19ab"). InnerVolumeSpecName "kube-api-access-lh2qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.114496 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh2qt\" (UniqueName: \"kubernetes.io/projected/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab-kube-api-access-lh2qt\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.500385 4983 generic.go:334] "Generic (PLEG): container finished" podID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" containerID="fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233" exitCode=0 Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.500425 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8w9dd" event={"ID":"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab","Type":"ContainerDied","Data":"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233"} Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.500465 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8w9dd" event={"ID":"e0f87abe-ea86-4f2b-be2f-4211bd4c19ab","Type":"ContainerDied","Data":"3b4183c71e4a59959a56a12584ad46d12c23ffc380617955c5e6a90fd59cb9bb"} Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.500482 4983 scope.go:117] "RemoveContainer" containerID="fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.500495 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8w9dd" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.501800 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-fxrbl" event={"ID":"a927379f-96be-4580-ad23-d807db431881","Type":"ContainerStarted","Data":"0f5f8064eceb17140d2f3dca1ef8c656fbce35abc5c6d6142f16b8f15bd66a4c"} Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.518119 4983 scope.go:117] "RemoveContainer" containerID="fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233" Oct 01 09:22:03 crc kubenswrapper[4983]: E1001 09:22:03.518541 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233\": container with ID starting with fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233 not found: ID does not exist" containerID="fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.518577 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233"} err="failed to get container status \"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233\": rpc error: code = NotFound desc = could not find container \"fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233\": container with ID starting with fa6ca02dc7da42d45499ddf07cae03d28b8bd7cd3568064b623b5deb5c801233 not found: ID does not exist" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.519531 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-fxrbl" podStartSLOduration=2.117371039 podStartE2EDuration="2.519522253s" podCreationTimestamp="2025-10-01 09:22:01 +0000 UTC" firstStartedPulling="2025-10-01 09:22:02.004595626 +0000 UTC m=+749.993824423" lastFinishedPulling="2025-10-01 09:22:02.40674684 +0000 UTC m=+750.395975637" observedRunningTime="2025-10-01 09:22:03.518799507 +0000 UTC m=+751.508028304" watchObservedRunningTime="2025-10-01 09:22:03.519522253 +0000 UTC m=+751.508751050" Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.535941 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:22:03 crc kubenswrapper[4983]: I1001 09:22:03.539914 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-8w9dd"] Oct 01 09:22:04 crc kubenswrapper[4983]: I1001 09:22:04.721158 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" path="/var/lib/kubelet/pods/e0f87abe-ea86-4f2b-be2f-4211bd4c19ab/volumes" Oct 01 09:22:06 crc kubenswrapper[4983]: I1001 09:22:06.881052 4983 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:22:11 crc kubenswrapper[4983]: I1001 09:22:11.774908 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:11 crc kubenswrapper[4983]: I1001 09:22:11.775470 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:11 crc kubenswrapper[4983]: I1001 09:22:11.802587 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:12 crc kubenswrapper[4983]: I1001 09:22:12.573006 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-fxrbl" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.082990 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn"] Oct 01 09:22:14 crc kubenswrapper[4983]: E1001 09:22:14.083552 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" containerName="registry-server" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.083569 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" containerName="registry-server" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.083707 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f87abe-ea86-4f2b-be2f-4211bd4c19ab" containerName="registry-server" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.084621 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.086840 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.096877 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn"] Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.144619 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-645c4\" (UniqueName: \"kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.144665 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.144713 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.245595 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.245690 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-645c4\" (UniqueName: \"kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.245713 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.246118 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.246165 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.264867 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-645c4\" (UniqueName: \"kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4\") pod \"e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.406258 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:14 crc kubenswrapper[4983]: I1001 09:22:14.867045 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn"] Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.582725 4983 generic.go:334] "Generic (PLEG): container finished" podID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerID="de50ad886235dee2ca23982cc83047c8146be402e020c35e2845db55f7c9fe58" exitCode=0 Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.582829 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" event={"ID":"756c9c8b-3d0a-4179-b980-aca2f443ed3f","Type":"ContainerDied","Data":"de50ad886235dee2ca23982cc83047c8146be402e020c35e2845db55f7c9fe58"} Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.583060 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" event={"ID":"756c9c8b-3d0a-4179-b980-aca2f443ed3f","Type":"ContainerStarted","Data":"0e87b6a71af8d3a54a0c817c97c0be552f900b0fe6a964e25cb2cff651fa8311"} Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.857930 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.859198 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.880787 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.971089 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.971168 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:15 crc kubenswrapper[4983]: I1001 09:22:15.971257 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jn8r\" (UniqueName: \"kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.072927 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.073011 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.073084 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jn8r\" (UniqueName: \"kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.073444 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.073444 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.096975 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jn8r\" (UniqueName: \"kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r\") pod \"redhat-marketplace-sg2v5\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.194499 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.587108 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.589821 4983 generic.go:334] "Generic (PLEG): container finished" podID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerID="8b5108f328816879d4e27c411127665699d515b01d243a82e5c2045758517336" exitCode=0 Oct 01 09:22:16 crc kubenswrapper[4983]: I1001 09:22:16.589846 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" event={"ID":"756c9c8b-3d0a-4179-b980-aca2f443ed3f","Type":"ContainerDied","Data":"8b5108f328816879d4e27c411127665699d515b01d243a82e5c2045758517336"} Oct 01 09:22:16 crc kubenswrapper[4983]: W1001 09:22:16.603681 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9e028b4_985a_4730_b31b_a00efb1e6517.slice/crio-108bdc0df19a764ffd2b1393ee2ed7f22f1085613c1a861759ca53241f948a5e WatchSource:0}: Error finding container 108bdc0df19a764ffd2b1393ee2ed7f22f1085613c1a861759ca53241f948a5e: Status 404 returned error can't find the container with id 108bdc0df19a764ffd2b1393ee2ed7f22f1085613c1a861759ca53241f948a5e Oct 01 09:22:17 crc kubenswrapper[4983]: I1001 09:22:17.603500 4983 generic.go:334] "Generic (PLEG): container finished" podID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerID="c4f7d2affc4fc0d126a971e3d761bda87a8dff4353d2da3c067d6d3e5384b21a" exitCode=0 Oct 01 09:22:17 crc kubenswrapper[4983]: I1001 09:22:17.603604 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" event={"ID":"756c9c8b-3d0a-4179-b980-aca2f443ed3f","Type":"ContainerDied","Data":"c4f7d2affc4fc0d126a971e3d761bda87a8dff4353d2da3c067d6d3e5384b21a"} Oct 01 09:22:17 crc kubenswrapper[4983]: I1001 09:22:17.607272 4983 generic.go:334] "Generic (PLEG): container finished" podID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerID="55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2" exitCode=0 Oct 01 09:22:17 crc kubenswrapper[4983]: I1001 09:22:17.607523 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerDied","Data":"55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2"} Oct 01 09:22:17 crc kubenswrapper[4983]: I1001 09:22:17.607725 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerStarted","Data":"108bdc0df19a764ffd2b1393ee2ed7f22f1085613c1a861759ca53241f948a5e"} Oct 01 09:22:18 crc kubenswrapper[4983]: I1001 09:22:18.981338 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.117266 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle\") pod \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.117331 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-645c4\" (UniqueName: \"kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4\") pod \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.117371 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util\") pod \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\" (UID: \"756c9c8b-3d0a-4179-b980-aca2f443ed3f\") " Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.118901 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle" (OuterVolumeSpecName: "bundle") pod "756c9c8b-3d0a-4179-b980-aca2f443ed3f" (UID: "756c9c8b-3d0a-4179-b980-aca2f443ed3f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.126454 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4" (OuterVolumeSpecName: "kube-api-access-645c4") pod "756c9c8b-3d0a-4179-b980-aca2f443ed3f" (UID: "756c9c8b-3d0a-4179-b980-aca2f443ed3f"). InnerVolumeSpecName "kube-api-access-645c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.151491 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util" (OuterVolumeSpecName: "util") pod "756c9c8b-3d0a-4179-b980-aca2f443ed3f" (UID: "756c9c8b-3d0a-4179-b980-aca2f443ed3f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.219899 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.219965 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-645c4\" (UniqueName: \"kubernetes.io/projected/756c9c8b-3d0a-4179-b980-aca2f443ed3f-kube-api-access-645c4\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.219995 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/756c9c8b-3d0a-4179-b980-aca2f443ed3f-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.630529 4983 generic.go:334] "Generic (PLEG): container finished" podID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerID="7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586" exitCode=0 Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.630605 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerDied","Data":"7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586"} Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.635035 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" event={"ID":"756c9c8b-3d0a-4179-b980-aca2f443ed3f","Type":"ContainerDied","Data":"0e87b6a71af8d3a54a0c817c97c0be552f900b0fe6a964e25cb2cff651fa8311"} Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.635084 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e87b6a71af8d3a54a0c817c97c0be552f900b0fe6a964e25cb2cff651fa8311" Oct 01 09:22:19 crc kubenswrapper[4983]: I1001 09:22:19.635190 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn" Oct 01 09:22:20 crc kubenswrapper[4983]: I1001 09:22:20.642769 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerStarted","Data":"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd"} Oct 01 09:22:20 crc kubenswrapper[4983]: I1001 09:22:20.661544 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sg2v5" podStartSLOduration=2.919971311 podStartE2EDuration="5.661525885s" podCreationTimestamp="2025-10-01 09:22:15 +0000 UTC" firstStartedPulling="2025-10-01 09:22:17.609343444 +0000 UTC m=+765.598572241" lastFinishedPulling="2025-10-01 09:22:20.350898018 +0000 UTC m=+768.340126815" observedRunningTime="2025-10-01 09:22:20.658077976 +0000 UTC m=+768.647306793" watchObservedRunningTime="2025-10-01 09:22:20.661525885 +0000 UTC m=+768.650754682" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.195384 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.195895 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.235716 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.595349 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6"] Oct 01 09:22:26 crc kubenswrapper[4983]: E1001 09:22:26.595792 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="util" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.595818 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="util" Oct 01 09:22:26 crc kubenswrapper[4983]: E1001 09:22:26.595833 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="pull" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.595839 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="pull" Oct 01 09:22:26 crc kubenswrapper[4983]: E1001 09:22:26.595854 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="extract" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.595860 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="extract" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.595952 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="756c9c8b-3d0a-4179-b980-aca2f443ed3f" containerName="extract" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.596495 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.598210 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ndnxk" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.598216 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.616839 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6"] Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.714800 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-apiservice-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.714911 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpkqr\" (UniqueName: \"kubernetes.io/projected/37538e82-453e-4a0a-858b-9929a1474605-kube-api-access-vpkqr\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.714974 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-webhook-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.722893 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.816328 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpkqr\" (UniqueName: \"kubernetes.io/projected/37538e82-453e-4a0a-858b-9929a1474605-kube-api-access-vpkqr\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.816417 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-webhook-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.816461 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-apiservice-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.821302 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-apiservice-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.821302 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37538e82-453e-4a0a-858b-9929a1474605-webhook-cert\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.832427 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpkqr\" (UniqueName: \"kubernetes.io/projected/37538e82-453e-4a0a-858b-9929a1474605-kube-api-access-vpkqr\") pod \"infra-operator-controller-manager-78d99444df-4zdf6\" (UID: \"37538e82-453e-4a0a-858b-9929a1474605\") " pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:26 crc kubenswrapper[4983]: I1001 09:22:26.916155 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:27 crc kubenswrapper[4983]: I1001 09:22:27.293692 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6"] Oct 01 09:22:27 crc kubenswrapper[4983]: W1001 09:22:27.303329 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37538e82_453e_4a0a_858b_9929a1474605.slice/crio-6185b9d248024a5ce4953b778134a381e2b9c4a46cd1024028ae753e31dcc2be WatchSource:0}: Error finding container 6185b9d248024a5ce4953b778134a381e2b9c4a46cd1024028ae753e31dcc2be: Status 404 returned error can't find the container with id 6185b9d248024a5ce4953b778134a381e2b9c4a46cd1024028ae753e31dcc2be Oct 01 09:22:27 crc kubenswrapper[4983]: I1001 09:22:27.683826 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" event={"ID":"37538e82-453e-4a0a-858b-9929a1474605","Type":"ContainerStarted","Data":"6185b9d248024a5ce4953b778134a381e2b9c4a46cd1024028ae753e31dcc2be"} Oct 01 09:22:29 crc kubenswrapper[4983]: I1001 09:22:29.698009 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" event={"ID":"37538e82-453e-4a0a-858b-9929a1474605","Type":"ContainerStarted","Data":"23c4e3ebc4762ec367f0baa887982d3c0440bff5b9b7b7e8374397a4bb57949a"} Oct 01 09:22:29 crc kubenswrapper[4983]: I1001 09:22:29.698568 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" event={"ID":"37538e82-453e-4a0a-858b-9929a1474605","Type":"ContainerStarted","Data":"7e6d7d3a7e6f6e29254df31f694534a7fad230bc053998c38ebd6a606ae96e21"} Oct 01 09:22:29 crc kubenswrapper[4983]: I1001 09:22:29.698612 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:29 crc kubenswrapper[4983]: I1001 09:22:29.720091 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" podStartSLOduration=1.797988761 podStartE2EDuration="3.720076147s" podCreationTimestamp="2025-10-01 09:22:26 +0000 UTC" firstStartedPulling="2025-10-01 09:22:27.30585955 +0000 UTC m=+775.295088357" lastFinishedPulling="2025-10-01 09:22:29.227946956 +0000 UTC m=+777.217175743" observedRunningTime="2025-10-01 09:22:29.71763953 +0000 UTC m=+777.706868327" watchObservedRunningTime="2025-10-01 09:22:29.720076147 +0000 UTC m=+777.709304944" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.045501 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.045732 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sg2v5" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="registry-server" containerID="cri-o://6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd" gracePeriod=2 Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.490669 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.560966 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jn8r\" (UniqueName: \"kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r\") pod \"d9e028b4-985a-4730-b31b-a00efb1e6517\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.561019 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities\") pod \"d9e028b4-985a-4730-b31b-a00efb1e6517\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.561038 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content\") pod \"d9e028b4-985a-4730-b31b-a00efb1e6517\" (UID: \"d9e028b4-985a-4730-b31b-a00efb1e6517\") " Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.562030 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities" (OuterVolumeSpecName: "utilities") pod "d9e028b4-985a-4730-b31b-a00efb1e6517" (UID: "d9e028b4-985a-4730-b31b-a00efb1e6517"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.566430 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r" (OuterVolumeSpecName: "kube-api-access-5jn8r") pod "d9e028b4-985a-4730-b31b-a00efb1e6517" (UID: "d9e028b4-985a-4730-b31b-a00efb1e6517"). InnerVolumeSpecName "kube-api-access-5jn8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.571762 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9e028b4-985a-4730-b31b-a00efb1e6517" (UID: "d9e028b4-985a-4730-b31b-a00efb1e6517"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.663309 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jn8r\" (UniqueName: \"kubernetes.io/projected/d9e028b4-985a-4730-b31b-a00efb1e6517-kube-api-access-5jn8r\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.663407 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.663436 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9e028b4-985a-4730-b31b-a00efb1e6517-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.707614 4983 generic.go:334] "Generic (PLEG): container finished" podID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerID="6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd" exitCode=0 Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.707653 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerDied","Data":"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd"} Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.707698 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg2v5" event={"ID":"d9e028b4-985a-4730-b31b-a00efb1e6517","Type":"ContainerDied","Data":"108bdc0df19a764ffd2b1393ee2ed7f22f1085613c1a861759ca53241f948a5e"} Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.707717 4983 scope.go:117] "RemoveContainer" containerID="6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.707758 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg2v5" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.711681 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.712144 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="extract-content" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.712181 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="extract-content" Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.712235 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="registry-server" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.712255 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="registry-server" Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.712282 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="extract-utilities" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.712300 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="extract-utilities" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.712521 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" containerName="registry-server" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.713959 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.718593 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.718684 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.718749 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.718800 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.718981 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.720119 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-b9zg9" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.730723 4983 scope.go:117] "RemoveContainer" containerID="7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.730936 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.753739 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.755871 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.769142 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.771767 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.777150 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.793902 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.799936 4983 scope.go:117] "RemoveContainer" containerID="55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.816091 4983 scope.go:117] "RemoveContainer" containerID="6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd" Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.816595 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd\": container with ID starting with 6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd not found: ID does not exist" containerID="6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.816626 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd"} err="failed to get container status \"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd\": rpc error: code = NotFound desc = could not find container \"6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd\": container with ID starting with 6ea4a8f85b7861dde3095a377a98af901d22bf89f1fb0e4e3215f56157bcd9bd not found: ID does not exist" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.816647 4983 scope.go:117] "RemoveContainer" containerID="7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586" Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.816963 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586\": container with ID starting with 7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586 not found: ID does not exist" containerID="7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.816987 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586"} err="failed to get container status \"7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586\": rpc error: code = NotFound desc = could not find container \"7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586\": container with ID starting with 7626475fe97b33cd49adcc48affe2e2cf1ddc2a23db28b41f51c0a1cfefe8586 not found: ID does not exist" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.817000 4983 scope.go:117] "RemoveContainer" containerID="55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2" Oct 01 09:22:30 crc kubenswrapper[4983]: E1001 09:22:30.817204 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2\": container with ID starting with 55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2 not found: ID does not exist" containerID="55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.817221 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2"} err="failed to get container status \"55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2\": rpc error: code = NotFound desc = could not find container \"55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2\": container with ID starting with 55040e3f2a8aab443087114a0cd36614e26c1ece41652de447875b028aef3ab2 not found: ID does not exist" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.823842 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.828147 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg2v5"] Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866166 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-default\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866240 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866364 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0cd75293-88f6-41e1-8d95-5b89e5a40f03-secrets\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866418 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfpgw\" (UniqueName: \"kubernetes.io/projected/f6110242-9584-4645-909c-514c48956b2d-kube-api-access-mfpgw\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866454 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ce45f506-705c-4232-a029-bd829ad6cae9-secrets\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866479 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866534 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-kolla-config\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866558 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5h7c\" (UniqueName: \"kubernetes.io/projected/ce45f506-705c-4232-a029-bd829ad6cae9-kube-api-access-p5h7c\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866578 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866594 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866615 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866640 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866665 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkd6c\" (UniqueName: \"kubernetes.io/projected/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kube-api-access-fkd6c\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866702 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-default\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866755 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866789 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-config-data-default\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866859 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f6110242-9584-4645-909c-514c48956b2d-secrets\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866884 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kolla-config\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866939 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-kolla-config\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.866965 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.867022 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6110242-9584-4645-909c-514c48956b2d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.968127 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6110242-9584-4645-909c-514c48956b2d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.968586 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-default\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.968528 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6110242-9584-4645-909c-514c48956b2d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969209 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.968608 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969378 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0cd75293-88f6-41e1-8d95-5b89e5a40f03-secrets\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969420 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfpgw\" (UniqueName: \"kubernetes.io/projected/f6110242-9584-4645-909c-514c48956b2d-kube-api-access-mfpgw\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969423 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-default\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969456 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ce45f506-705c-4232-a029-bd829ad6cae9-secrets\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969487 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969561 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-kolla-config\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969593 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5h7c\" (UniqueName: \"kubernetes.io/projected/ce45f506-705c-4232-a029-bd829ad6cae9-kube-api-access-p5h7c\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969621 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969639 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969662 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969687 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969710 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkd6c\" (UniqueName: \"kubernetes.io/projected/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kube-api-access-fkd6c\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969735 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-default\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969770 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969799 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-config-data-default\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969864 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kolla-config\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969886 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f6110242-9584-4645-909c-514c48956b2d-secrets\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969924 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-kolla-config\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.969951 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.970318 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.970858 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-kolla-config\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.970997 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-config-data-default\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971234 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971314 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971490 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971581 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kolla-config\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971675 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.971946 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6110242-9584-4645-909c-514c48956b2d-config-data-default\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.972237 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-kolla-config\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.972472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ce45f506-705c-4232-a029-bd829ad6cae9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.973166 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce45f506-705c-4232-a029-bd829ad6cae9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.974900 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0cd75293-88f6-41e1-8d95-5b89e5a40f03-secrets\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.975195 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ce45f506-705c-4232-a029-bd829ad6cae9-secrets\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.975554 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f6110242-9584-4645-909c-514c48956b2d-secrets\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.986616 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfpgw\" (UniqueName: \"kubernetes.io/projected/f6110242-9584-4645-909c-514c48956b2d-kube-api-access-mfpgw\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.987784 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"f6110242-9584-4645-909c-514c48956b2d\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.987877 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5h7c\" (UniqueName: \"kubernetes.io/projected/ce45f506-705c-4232-a029-bd829ad6cae9-kube-api-access-p5h7c\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.987969 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.989905 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"ce45f506-705c-4232-a029-bd829ad6cae9\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:30 crc kubenswrapper[4983]: I1001 09:22:30.998113 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkd6c\" (UniqueName: \"kubernetes.io/projected/0cd75293-88f6-41e1-8d95-5b89e5a40f03-kube-api-access-fkd6c\") pod \"openstack-galera-2\" (UID: \"0cd75293-88f6-41e1-8d95-5b89e5a40f03\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.085497 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.143604 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.150975 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.516872 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 01 09:22:31 crc kubenswrapper[4983]: W1001 09:22:31.520244 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce45f506_705c_4232_a029_bd829ad6cae9.slice/crio-41560d8dedaa2c484e8d5a133ebc6b5bacdfb0008550fe8bb57e9121c0b5763e WatchSource:0}: Error finding container 41560d8dedaa2c484e8d5a133ebc6b5bacdfb0008550fe8bb57e9121c0b5763e: Status 404 returned error can't find the container with id 41560d8dedaa2c484e8d5a133ebc6b5bacdfb0008550fe8bb57e9121c0b5763e Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.594771 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 01 09:22:31 crc kubenswrapper[4983]: W1001 09:22:31.601847 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6110242_9584_4645_909c_514c48956b2d.slice/crio-61ba976164cdf1aa889be4d3e8a0219ff911ce490255afeafe310d7e0a188d0b WatchSource:0}: Error finding container 61ba976164cdf1aa889be4d3e8a0219ff911ce490255afeafe310d7e0a188d0b: Status 404 returned error can't find the container with id 61ba976164cdf1aa889be4d3e8a0219ff911ce490255afeafe310d7e0a188d0b Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.603903 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 01 09:22:31 crc kubenswrapper[4983]: W1001 09:22:31.606138 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd75293_88f6_41e1_8d95_5b89e5a40f03.slice/crio-dacf5b377176de900d5fcaf1e71c8321a89a159169021d8d57d90aed4a0a7b75 WatchSource:0}: Error finding container dacf5b377176de900d5fcaf1e71c8321a89a159169021d8d57d90aed4a0a7b75: Status 404 returned error can't find the container with id dacf5b377176de900d5fcaf1e71c8321a89a159169021d8d57d90aed4a0a7b75 Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.714973 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"f6110242-9584-4645-909c-514c48956b2d","Type":"ContainerStarted","Data":"61ba976164cdf1aa889be4d3e8a0219ff911ce490255afeafe310d7e0a188d0b"} Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.715935 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"0cd75293-88f6-41e1-8d95-5b89e5a40f03","Type":"ContainerStarted","Data":"dacf5b377176de900d5fcaf1e71c8321a89a159169021d8d57d90aed4a0a7b75"} Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.717303 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"ce45f506-705c-4232-a029-bd829ad6cae9","Type":"ContainerStarted","Data":"41560d8dedaa2c484e8d5a133ebc6b5bacdfb0008550fe8bb57e9121c0b5763e"} Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.960350 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:22:31 crc kubenswrapper[4983]: I1001 09:22:31.960401 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.062074 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.064233 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.073519 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.188020 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.188088 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.188112 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-986tr\" (UniqueName: \"kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.289101 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.289147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-986tr\" (UniqueName: \"kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.289225 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.289708 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.289715 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.308933 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-986tr\" (UniqueName: \"kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr\") pod \"community-operators-srl8q\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.385668 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.746277 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9e028b4-985a-4730-b31b-a00efb1e6517" path="/var/lib/kubelet/pods/d9e028b4-985a-4730-b31b-a00efb1e6517/volumes" Oct 01 09:22:32 crc kubenswrapper[4983]: I1001 09:22:32.865231 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:22:32 crc kubenswrapper[4983]: W1001 09:22:32.871596 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d583ae4_a287_4fcf_b132_b00ca3a797f6.slice/crio-d440ac0d0275fa3c7b556c6492773df2b89f4f1e8696c857ed2fd5892bb6bf4d WatchSource:0}: Error finding container d440ac0d0275fa3c7b556c6492773df2b89f4f1e8696c857ed2fd5892bb6bf4d: Status 404 returned error can't find the container with id d440ac0d0275fa3c7b556c6492773df2b89f4f1e8696c857ed2fd5892bb6bf4d Oct 01 09:22:33 crc kubenswrapper[4983]: I1001 09:22:33.744964 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerID="0cbce8fa7669d60f292058e86e0f01b8f278bd854f9fd305db8549106d894f95" exitCode=0 Oct 01 09:22:33 crc kubenswrapper[4983]: I1001 09:22:33.745037 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerDied","Data":"0cbce8fa7669d60f292058e86e0f01b8f278bd854f9fd305db8549106d894f95"} Oct 01 09:22:33 crc kubenswrapper[4983]: I1001 09:22:33.745532 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerStarted","Data":"d440ac0d0275fa3c7b556c6492773df2b89f4f1e8696c857ed2fd5892bb6bf4d"} Oct 01 09:22:36 crc kubenswrapper[4983]: I1001 09:22:36.921245 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d99444df-4zdf6" Oct 01 09:22:38 crc kubenswrapper[4983]: I1001 09:22:38.773380 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"ce45f506-705c-4232-a029-bd829ad6cae9","Type":"ContainerStarted","Data":"c873d9990ffc18cde8abeaff9531b90bcde6e2eba42c8ae708a00579a55ebc48"} Oct 01 09:22:38 crc kubenswrapper[4983]: I1001 09:22:38.776348 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerStarted","Data":"cc6e8a8f61ad2b986e30e734f43eb1141096c466e55890262940b1deb35af489"} Oct 01 09:22:38 crc kubenswrapper[4983]: I1001 09:22:38.777793 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"f6110242-9584-4645-909c-514c48956b2d","Type":"ContainerStarted","Data":"4e113b44b8d186997479c5814d856d507ae98bf5ac0bbb1e35d38559f20a32fb"} Oct 01 09:22:38 crc kubenswrapper[4983]: I1001 09:22:38.779998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"0cd75293-88f6-41e1-8d95-5b89e5a40f03","Type":"ContainerStarted","Data":"2e2370a253ce829b67b87b362af14bc2714ed42e9a8f65256976e566442d1a2f"} Oct 01 09:22:39 crc kubenswrapper[4983]: I1001 09:22:39.789931 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerID="cc6e8a8f61ad2b986e30e734f43eb1141096c466e55890262940b1deb35af489" exitCode=0 Oct 01 09:22:39 crc kubenswrapper[4983]: I1001 09:22:39.789989 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerDied","Data":"cc6e8a8f61ad2b986e30e734f43eb1141096c466e55890262940b1deb35af489"} Oct 01 09:22:40 crc kubenswrapper[4983]: I1001 09:22:40.797156 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerStarted","Data":"97b9c1959f95ff2c05f626106c4be42f7f9d664048edad244bd2419ee056b8fd"} Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.054454 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-srl8q" podStartSLOduration=3.214713561 podStartE2EDuration="10.054439058s" podCreationTimestamp="2025-10-01 09:22:32 +0000 UTC" firstStartedPulling="2025-10-01 09:22:33.747154775 +0000 UTC m=+781.736383582" lastFinishedPulling="2025-10-01 09:22:40.586880282 +0000 UTC m=+788.576109079" observedRunningTime="2025-10-01 09:22:40.813988831 +0000 UTC m=+788.803217628" watchObservedRunningTime="2025-10-01 09:22:42.054439058 +0000 UTC m=+790.043667855" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.056319 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-xhlpw"] Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.057020 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.059732 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-wvld5" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.065554 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-xhlpw"] Oct 01 09:22:42 crc kubenswrapper[4983]: E1001 09:22:42.154243 4983 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6110242_9584_4645_909c_514c48956b2d.slice/crio-4e113b44b8d186997479c5814d856d507ae98bf5ac0bbb1e35d38559f20a32fb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6110242_9584_4645_909c_514c48956b2d.slice/crio-conmon-4e113b44b8d186997479c5814d856d507ae98bf5ac0bbb1e35d38559f20a32fb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce45f506_705c_4232_a029_bd829ad6cae9.slice/crio-c873d9990ffc18cde8abeaff9531b90bcde6e2eba42c8ae708a00579a55ebc48.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.233250 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl25j\" (UniqueName: \"kubernetes.io/projected/3465433a-f044-4239-b1d6-bab1711de522-kube-api-access-bl25j\") pod \"rabbitmq-cluster-operator-index-xhlpw\" (UID: \"3465433a-f044-4239-b1d6-bab1711de522\") " pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.334643 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl25j\" (UniqueName: \"kubernetes.io/projected/3465433a-f044-4239-b1d6-bab1711de522-kube-api-access-bl25j\") pod \"rabbitmq-cluster-operator-index-xhlpw\" (UID: \"3465433a-f044-4239-b1d6-bab1711de522\") " pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.353480 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl25j\" (UniqueName: \"kubernetes.io/projected/3465433a-f044-4239-b1d6-bab1711de522-kube-api-access-bl25j\") pod \"rabbitmq-cluster-operator-index-xhlpw\" (UID: \"3465433a-f044-4239-b1d6-bab1711de522\") " pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.371241 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.386470 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.386581 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.762952 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-xhlpw"] Oct 01 09:22:42 crc kubenswrapper[4983]: W1001 09:22:42.768540 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3465433a_f044_4239_b1d6_bab1711de522.slice/crio-ad18d235d23f0ace4569cf4925183ec2c28dc511f3e0c13f85f0380003bc79a0 WatchSource:0}: Error finding container ad18d235d23f0ace4569cf4925183ec2c28dc511f3e0c13f85f0380003bc79a0: Status 404 returned error can't find the container with id ad18d235d23f0ace4569cf4925183ec2c28dc511f3e0c13f85f0380003bc79a0 Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.809916 4983 generic.go:334] "Generic (PLEG): container finished" podID="ce45f506-705c-4232-a029-bd829ad6cae9" containerID="c873d9990ffc18cde8abeaff9531b90bcde6e2eba42c8ae708a00579a55ebc48" exitCode=0 Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.809997 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"ce45f506-705c-4232-a029-bd829ad6cae9","Type":"ContainerDied","Data":"c873d9990ffc18cde8abeaff9531b90bcde6e2eba42c8ae708a00579a55ebc48"} Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.811983 4983 generic.go:334] "Generic (PLEG): container finished" podID="f6110242-9584-4645-909c-514c48956b2d" containerID="4e113b44b8d186997479c5814d856d507ae98bf5ac0bbb1e35d38559f20a32fb" exitCode=0 Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.812055 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"f6110242-9584-4645-909c-514c48956b2d","Type":"ContainerDied","Data":"4e113b44b8d186997479c5814d856d507ae98bf5ac0bbb1e35d38559f20a32fb"} Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.813131 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" event={"ID":"3465433a-f044-4239-b1d6-bab1711de522","Type":"ContainerStarted","Data":"ad18d235d23f0ace4569cf4925183ec2c28dc511f3e0c13f85f0380003bc79a0"} Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.814494 4983 generic.go:334] "Generic (PLEG): container finished" podID="0cd75293-88f6-41e1-8d95-5b89e5a40f03" containerID="2e2370a253ce829b67b87b362af14bc2714ed42e9a8f65256976e566442d1a2f" exitCode=0 Oct 01 09:22:42 crc kubenswrapper[4983]: I1001 09:22:42.814571 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"0cd75293-88f6-41e1-8d95-5b89e5a40f03","Type":"ContainerDied","Data":"2e2370a253ce829b67b87b362af14bc2714ed42e9a8f65256976e566442d1a2f"} Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.429319 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-srl8q" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="registry-server" probeResult="failure" output=< Oct 01 09:22:43 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Oct 01 09:22:43 crc kubenswrapper[4983]: > Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.824960 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"0cd75293-88f6-41e1-8d95-5b89e5a40f03","Type":"ContainerStarted","Data":"918b432a44f2e78479ebac40482030525d227baccc2abb73462b12ec0152af12"} Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.827383 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"ce45f506-705c-4232-a029-bd829ad6cae9","Type":"ContainerStarted","Data":"a75a7a8ed91aff723e924d7304b840f16ddbc3523737640823f5d973e2bfad88"} Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.829577 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"f6110242-9584-4645-909c-514c48956b2d","Type":"ContainerStarted","Data":"bad88b82eca648f2b7940107c8bbc6b20dea2269295fba85a77370f85e9a6d1c"} Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.850216 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.020270137 podStartE2EDuration="14.850194106s" podCreationTimestamp="2025-10-01 09:22:29 +0000 UTC" firstStartedPulling="2025-10-01 09:22:31.609281615 +0000 UTC m=+779.598510412" lastFinishedPulling="2025-10-01 09:22:38.439205594 +0000 UTC m=+786.428434381" observedRunningTime="2025-10-01 09:22:43.844526721 +0000 UTC m=+791.833755538" watchObservedRunningTime="2025-10-01 09:22:43.850194106 +0000 UTC m=+791.839422913" Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.874472 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=7.917637522 podStartE2EDuration="14.874455989s" podCreationTimestamp="2025-10-01 09:22:29 +0000 UTC" firstStartedPulling="2025-10-01 09:22:31.522290478 +0000 UTC m=+779.511519275" lastFinishedPulling="2025-10-01 09:22:38.479108945 +0000 UTC m=+786.468337742" observedRunningTime="2025-10-01 09:22:43.872248839 +0000 UTC m=+791.861477636" watchObservedRunningTime="2025-10-01 09:22:43.874455989 +0000 UTC m=+791.863684786" Oct 01 09:22:43 crc kubenswrapper[4983]: I1001 09:22:43.889675 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=8.053954768 podStartE2EDuration="14.889656874s" podCreationTimestamp="2025-10-01 09:22:29 +0000 UTC" firstStartedPulling="2025-10-01 09:22:31.60391319 +0000 UTC m=+779.593141987" lastFinishedPulling="2025-10-01 09:22:38.439615296 +0000 UTC m=+786.428844093" observedRunningTime="2025-10-01 09:22:43.888493972 +0000 UTC m=+791.877722769" watchObservedRunningTime="2025-10-01 09:22:43.889656874 +0000 UTC m=+791.878885671" Oct 01 09:22:45 crc kubenswrapper[4983]: E1001 09:22:45.007231 4983 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.194:34092->38.102.83.194:44187: write tcp 38.102.83.194:34092->38.102.83.194:44187: write: broken pipe Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.021781 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.022852 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.025520 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.030947 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.035174 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-7gh6w" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.126839 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-config-data\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.126899 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fngn2\" (UniqueName: \"kubernetes.io/projected/d44fa8ad-8a99-41d6-8dba-11911de2c065-kube-api-access-fngn2\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.126986 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-kolla-config\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.227930 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-config-data\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.228026 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fngn2\" (UniqueName: \"kubernetes.io/projected/d44fa8ad-8a99-41d6-8dba-11911de2c065-kube-api-access-fngn2\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.228058 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-kolla-config\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.228914 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-config-data\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.228920 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d44fa8ad-8a99-41d6-8dba-11911de2c065-kolla-config\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.253863 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fngn2\" (UniqueName: \"kubernetes.io/projected/d44fa8ad-8a99-41d6-8dba-11911de2c065-kube-api-access-fngn2\") pod \"memcached-0\" (UID: \"d44fa8ad-8a99-41d6-8dba-11911de2c065\") " pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.337688 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.710865 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.859977 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" event={"ID":"3465433a-f044-4239-b1d6-bab1711de522","Type":"ContainerStarted","Data":"196ce49a20a871a2d26a0f5ec507992beeba7d200be49ce98d00a4b7c22d27a3"} Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.861522 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"d44fa8ad-8a99-41d6-8dba-11911de2c065","Type":"ContainerStarted","Data":"02d23efc04174c633e70eec3ba0dd0b3007666e1e0fccf504111c0f592e12ce0"} Oct 01 09:22:48 crc kubenswrapper[4983]: I1001 09:22:48.874248 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" podStartSLOduration=1.79836664 podStartE2EDuration="6.87423086s" podCreationTimestamp="2025-10-01 09:22:42 +0000 UTC" firstStartedPulling="2025-10-01 09:22:42.770342868 +0000 UTC m=+790.759571665" lastFinishedPulling="2025-10-01 09:22:47.846207088 +0000 UTC m=+795.835435885" observedRunningTime="2025-10-01 09:22:48.873727126 +0000 UTC m=+796.862955933" watchObservedRunningTime="2025-10-01 09:22:48.87423086 +0000 UTC m=+796.863459657" Oct 01 09:22:49 crc kubenswrapper[4983]: E1001 09:22:49.508761 4983 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.194:34134->38.102.83.194:44187: write tcp 38.102.83.194:34134->38.102.83.194:44187: write: connection reset by peer Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.086088 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.086883 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.144194 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.144259 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.152086 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.152134 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.885571 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"d44fa8ad-8a99-41d6-8dba-11911de2c065","Type":"ContainerStarted","Data":"ed81c5ef8e0f69d2b0c7b5bfba9fe34975aef81dfb2291bfe100aa61a98c3898"} Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.886172 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:51 crc kubenswrapper[4983]: I1001 09:22:51.903691 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=1.686283728 podStartE2EDuration="3.90365613s" podCreationTimestamp="2025-10-01 09:22:48 +0000 UTC" firstStartedPulling="2025-10-01 09:22:48.716460607 +0000 UTC m=+796.705689404" lastFinishedPulling="2025-10-01 09:22:50.933833009 +0000 UTC m=+798.923061806" observedRunningTime="2025-10-01 09:22:51.898844638 +0000 UTC m=+799.888073435" watchObservedRunningTime="2025-10-01 09:22:51.90365613 +0000 UTC m=+799.892884927" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.371856 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.371923 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.398390 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.427636 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.466483 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:52 crc kubenswrapper[4983]: I1001 09:22:52.913275 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-xhlpw" Oct 01 09:22:53 crc kubenswrapper[4983]: I1001 09:22:53.216189 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:53 crc kubenswrapper[4983]: I1001 09:22:53.260709 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Oct 01 09:22:53 crc kubenswrapper[4983]: E1001 09:22:53.432570 4983 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.194:49710->38.102.83.194:44187: write tcp 38.102.83.194:49710->38.102.83.194:44187: write: broken pipe Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.455732 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.457364 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.466057 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.543437 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ljtj\" (UniqueName: \"kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.543516 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.543636 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.644962 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.645268 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ljtj\" (UniqueName: \"kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.645378 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.645907 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.646039 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.683061 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ljtj\" (UniqueName: \"kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj\") pod \"certified-operators-fkcsd\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.724107 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj"] Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.725608 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.728219 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.750034 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj"] Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.785545 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.852554 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjp6f\" (UniqueName: \"kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.852718 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.852760 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.954173 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.954514 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.954545 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjp6f\" (UniqueName: \"kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.954708 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.955029 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:56 crc kubenswrapper[4983]: I1001 09:22:56.972716 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjp6f\" (UniqueName: \"kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.053971 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.201427 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.438382 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj"] Oct 01 09:22:57 crc kubenswrapper[4983]: W1001 09:22:57.444675 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5649915e_2f6c_4e93_8aba_9820825b4643.slice/crio-ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216 WatchSource:0}: Error finding container ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216: Status 404 returned error can't find the container with id ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216 Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.929497 4983 generic.go:334] "Generic (PLEG): container finished" podID="5649915e-2f6c-4e93-8aba-9820825b4643" containerID="489b0f5d8ad4b62d3bac2e7e294cc34244aa60c7d632cfe925eba21607a59a40" exitCode=0 Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.929592 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" event={"ID":"5649915e-2f6c-4e93-8aba-9820825b4643","Type":"ContainerDied","Data":"489b0f5d8ad4b62d3bac2e7e294cc34244aa60c7d632cfe925eba21607a59a40"} Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.930069 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" event={"ID":"5649915e-2f6c-4e93-8aba-9820825b4643","Type":"ContainerStarted","Data":"ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216"} Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.931897 4983 generic.go:334] "Generic (PLEG): container finished" podID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerID="dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791" exitCode=0 Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.931931 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerDied","Data":"dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791"} Oct 01 09:22:57 crc kubenswrapper[4983]: I1001 09:22:57.931952 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerStarted","Data":"67016667aa4c71997e0150d2071178c994e5f3c4894026cde3c1477c01ac90a7"} Oct 01 09:22:58 crc kubenswrapper[4983]: I1001 09:22:58.339433 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Oct 01 09:22:58 crc kubenswrapper[4983]: I1001 09:22:58.451687 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:22:58 crc kubenswrapper[4983]: I1001 09:22:58.452055 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-srl8q" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="registry-server" containerID="cri-o://97b9c1959f95ff2c05f626106c4be42f7f9d664048edad244bd2419ee056b8fd" gracePeriod=2 Oct 01 09:22:58 crc kubenswrapper[4983]: I1001 09:22:58.940102 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerID="97b9c1959f95ff2c05f626106c4be42f7f9d664048edad244bd2419ee056b8fd" exitCode=0 Oct 01 09:22:58 crc kubenswrapper[4983]: I1001 09:22:58.940171 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerDied","Data":"97b9c1959f95ff2c05f626106c4be42f7f9d664048edad244bd2419ee056b8fd"} Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.637885 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.688618 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content\") pod \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.688676 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities\") pod \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.688772 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-986tr\" (UniqueName: \"kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr\") pod \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\" (UID: \"5d583ae4-a287-4fcf-b132-b00ca3a797f6\") " Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.689773 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities" (OuterVolumeSpecName: "utilities") pod "5d583ae4-a287-4fcf-b132-b00ca3a797f6" (UID: "5d583ae4-a287-4fcf-b132-b00ca3a797f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.695098 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr" (OuterVolumeSpecName: "kube-api-access-986tr") pod "5d583ae4-a287-4fcf-b132-b00ca3a797f6" (UID: "5d583ae4-a287-4fcf-b132-b00ca3a797f6"). InnerVolumeSpecName "kube-api-access-986tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.742087 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d583ae4-a287-4fcf-b132-b00ca3a797f6" (UID: "5d583ae4-a287-4fcf-b132-b00ca3a797f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.790380 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-986tr\" (UniqueName: \"kubernetes.io/projected/5d583ae4-a287-4fcf-b132-b00ca3a797f6-kube-api-access-986tr\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.790425 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.790439 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d583ae4-a287-4fcf-b132-b00ca3a797f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.949289 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srl8q" event={"ID":"5d583ae4-a287-4fcf-b132-b00ca3a797f6","Type":"ContainerDied","Data":"d440ac0d0275fa3c7b556c6492773df2b89f4f1e8696c857ed2fd5892bb6bf4d"} Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.949311 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srl8q" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.949345 4983 scope.go:117] "RemoveContainer" containerID="97b9c1959f95ff2c05f626106c4be42f7f9d664048edad244bd2419ee056b8fd" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.959624 4983 generic.go:334] "Generic (PLEG): container finished" podID="5649915e-2f6c-4e93-8aba-9820825b4643" containerID="4fb475b2b00169df553259db6050ffc1e39eb20f84e6bc56af9582ea0e5fb1b6" exitCode=0 Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.959687 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" event={"ID":"5649915e-2f6c-4e93-8aba-9820825b4643","Type":"ContainerDied","Data":"4fb475b2b00169df553259db6050ffc1e39eb20f84e6bc56af9582ea0e5fb1b6"} Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.964895 4983 generic.go:334] "Generic (PLEG): container finished" podID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerID="9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418" exitCode=0 Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.964942 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerDied","Data":"9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418"} Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.974052 4983 scope.go:117] "RemoveContainer" containerID="cc6e8a8f61ad2b986e30e734f43eb1141096c466e55890262940b1deb35af489" Oct 01 09:22:59 crc kubenswrapper[4983]: I1001 09:22:59.994227 4983 scope.go:117] "RemoveContainer" containerID="0cbce8fa7669d60f292058e86e0f01b8f278bd854f9fd305db8549106d894f95" Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.018320 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.023618 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-srl8q"] Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.721520 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" path="/var/lib/kubelet/pods/5d583ae4-a287-4fcf-b132-b00ca3a797f6/volumes" Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.973319 4983 generic.go:334] "Generic (PLEG): container finished" podID="5649915e-2f6c-4e93-8aba-9820825b4643" containerID="b0f1393dbaaeb832b0f8d88d59dbdddf0f473235e74617ccc3ee7514a7633641" exitCode=0 Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.973413 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" event={"ID":"5649915e-2f6c-4e93-8aba-9820825b4643","Type":"ContainerDied","Data":"b0f1393dbaaeb832b0f8d88d59dbdddf0f473235e74617ccc3ee7514a7633641"} Oct 01 09:23:00 crc kubenswrapper[4983]: I1001 09:23:00.975271 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerStarted","Data":"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab"} Oct 01 09:23:01 crc kubenswrapper[4983]: I1001 09:23:01.014027 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fkcsd" podStartSLOduration=2.542199607 podStartE2EDuration="5.014010965s" podCreationTimestamp="2025-10-01 09:22:56 +0000 UTC" firstStartedPulling="2025-10-01 09:22:57.932940661 +0000 UTC m=+805.922169458" lastFinishedPulling="2025-10-01 09:23:00.404752019 +0000 UTC m=+808.393980816" observedRunningTime="2025-10-01 09:23:01.012280227 +0000 UTC m=+809.001509024" watchObservedRunningTime="2025-10-01 09:23:01.014010965 +0000 UTC m=+809.003239762" Oct 01 09:23:01 crc kubenswrapper[4983]: I1001 09:23:01.960870 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:23:01 crc kubenswrapper[4983]: I1001 09:23:01.961177 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.264777 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.323138 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle\") pod \"5649915e-2f6c-4e93-8aba-9820825b4643\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.323198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjp6f\" (UniqueName: \"kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f\") pod \"5649915e-2f6c-4e93-8aba-9820825b4643\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.323290 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util\") pod \"5649915e-2f6c-4e93-8aba-9820825b4643\" (UID: \"5649915e-2f6c-4e93-8aba-9820825b4643\") " Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.323889 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle" (OuterVolumeSpecName: "bundle") pod "5649915e-2f6c-4e93-8aba-9820825b4643" (UID: "5649915e-2f6c-4e93-8aba-9820825b4643"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.327919 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f" (OuterVolumeSpecName: "kube-api-access-sjp6f") pod "5649915e-2f6c-4e93-8aba-9820825b4643" (UID: "5649915e-2f6c-4e93-8aba-9820825b4643"). InnerVolumeSpecName "kube-api-access-sjp6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.333910 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util" (OuterVolumeSpecName: "util") pod "5649915e-2f6c-4e93-8aba-9820825b4643" (UID: "5649915e-2f6c-4e93-8aba-9820825b4643"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.424725 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.424758 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjp6f\" (UniqueName: \"kubernetes.io/projected/5649915e-2f6c-4e93-8aba-9820825b4643-kube-api-access-sjp6f\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.424770 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5649915e-2f6c-4e93-8aba-9820825b4643-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.991702 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" event={"ID":"5649915e-2f6c-4e93-8aba-9820825b4643","Type":"ContainerDied","Data":"ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216"} Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.992076 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe9e19eff75aaebfd6970ca76baea745b7a025809a9fad7263e0870f4d17216" Oct 01 09:23:02 crc kubenswrapper[4983]: I1001 09:23:02.991916 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj" Oct 01 09:23:04 crc kubenswrapper[4983]: I1001 09:23:04.453456 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:23:04 crc kubenswrapper[4983]: I1001 09:23:04.497482 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Oct 01 09:23:06 crc kubenswrapper[4983]: I1001 09:23:06.786040 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:06 crc kubenswrapper[4983]: I1001 09:23:06.786325 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:06 crc kubenswrapper[4983]: I1001 09:23:06.822222 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:07 crc kubenswrapper[4983]: I1001 09:23:07.051190 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:08 crc kubenswrapper[4983]: I1001 09:23:08.676430 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:23:08 crc kubenswrapper[4983]: I1001 09:23:08.752308 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.247378 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.247875 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fkcsd" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="registry-server" containerID="cri-o://36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab" gracePeriod=2 Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.640004 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.730976 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ljtj\" (UniqueName: \"kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj\") pod \"9c286f04-03dd-4138-a328-c5c44c09a7a1\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.731326 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content\") pod \"9c286f04-03dd-4138-a328-c5c44c09a7a1\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.731362 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities\") pod \"9c286f04-03dd-4138-a328-c5c44c09a7a1\" (UID: \"9c286f04-03dd-4138-a328-c5c44c09a7a1\") " Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.732425 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities" (OuterVolumeSpecName: "utilities") pod "9c286f04-03dd-4138-a328-c5c44c09a7a1" (UID: "9c286f04-03dd-4138-a328-c5c44c09a7a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.737982 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj" (OuterVolumeSpecName: "kube-api-access-9ljtj") pod "9c286f04-03dd-4138-a328-c5c44c09a7a1" (UID: "9c286f04-03dd-4138-a328-c5c44c09a7a1"). InnerVolumeSpecName "kube-api-access-9ljtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.773778 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c286f04-03dd-4138-a328-c5c44c09a7a1" (UID: "9c286f04-03dd-4138-a328-c5c44c09a7a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.834333 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ljtj\" (UniqueName: \"kubernetes.io/projected/9c286f04-03dd-4138-a328-c5c44c09a7a1-kube-api-access-9ljtj\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.834368 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:10 crc kubenswrapper[4983]: I1001 09:23:10.834380 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c286f04-03dd-4138-a328-c5c44c09a7a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.037928 4983 generic.go:334] "Generic (PLEG): container finished" podID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerID="36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab" exitCode=0 Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.037991 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerDied","Data":"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab"} Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.038006 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkcsd" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.038074 4983 scope.go:117] "RemoveContainer" containerID="36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.038057 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkcsd" event={"ID":"9c286f04-03dd-4138-a328-c5c44c09a7a1","Type":"ContainerDied","Data":"67016667aa4c71997e0150d2071178c994e5f3c4894026cde3c1477c01ac90a7"} Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.052345 4983 scope.go:117] "RemoveContainer" containerID="9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.067281 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.069198 4983 scope.go:117] "RemoveContainer" containerID="dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.073741 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fkcsd"] Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.093410 4983 scope.go:117] "RemoveContainer" containerID="36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab" Oct 01 09:23:11 crc kubenswrapper[4983]: E1001 09:23:11.094132 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab\": container with ID starting with 36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab not found: ID does not exist" containerID="36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.094205 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab"} err="failed to get container status \"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab\": rpc error: code = NotFound desc = could not find container \"36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab\": container with ID starting with 36a255802cf6bb55a90898c594071b9aca5e55fe335d4954d734f7366a56baab not found: ID does not exist" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.094241 4983 scope.go:117] "RemoveContainer" containerID="9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418" Oct 01 09:23:11 crc kubenswrapper[4983]: E1001 09:23:11.094880 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418\": container with ID starting with 9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418 not found: ID does not exist" containerID="9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.094928 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418"} err="failed to get container status \"9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418\": rpc error: code = NotFound desc = could not find container \"9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418\": container with ID starting with 9dd2bc9c68474900d14fb783fdeb2d1a59eb7e3210868d885f9b50159a671418 not found: ID does not exist" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.095006 4983 scope.go:117] "RemoveContainer" containerID="dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791" Oct 01 09:23:11 crc kubenswrapper[4983]: E1001 09:23:11.095462 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791\": container with ID starting with dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791 not found: ID does not exist" containerID="dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791" Oct 01 09:23:11 crc kubenswrapper[4983]: I1001 09:23:11.095493 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791"} err="failed to get container status \"dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791\": rpc error: code = NotFound desc = could not find container \"dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791\": container with ID starting with dbd4ab8201ab919c0f63dbee9bf0ba8365bcc68bf6abf0462902941eec67f791 not found: ID does not exist" Oct 01 09:23:12 crc kubenswrapper[4983]: I1001 09:23:12.724701 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" path="/var/lib/kubelet/pods/9c286f04-03dd-4138-a328-c5c44c09a7a1/volumes" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.055933 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056187 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="extract" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056198 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="extract" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056209 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056218 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056234 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="extract-content" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056243 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="extract-content" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056252 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="util" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056261 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="util" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056276 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="extract-utilities" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056284 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="extract-utilities" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056303 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056309 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056317 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="extract-content" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056322 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="extract-content" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056334 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="pull" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056340 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="pull" Oct 01 09:23:13 crc kubenswrapper[4983]: E1001 09:23:13.056349 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="extract-utilities" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056354 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="extract-utilities" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056456 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c286f04-03dd-4138-a328-c5c44c09a7a1" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056472 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5649915e-2f6c-4e93-8aba-9820825b4643" containerName="extract" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.056479 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d583ae4-a287-4fcf-b132-b00ca3a797f6" containerName="registry-server" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.057255 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.066713 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.160081 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.160415 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dq7d\" (UniqueName: \"kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.160545 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.262402 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.262912 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.263057 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dq7d\" (UniqueName: \"kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.263091 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.263485 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.280618 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dq7d\" (UniqueName: \"kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d\") pod \"redhat-operators-b6cmc\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.372472 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:13 crc kubenswrapper[4983]: I1001 09:23:13.836551 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:14 crc kubenswrapper[4983]: I1001 09:23:14.059923 4983 generic.go:334] "Generic (PLEG): container finished" podID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerID="3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a" exitCode=0 Oct 01 09:23:14 crc kubenswrapper[4983]: I1001 09:23:14.060033 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerDied","Data":"3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a"} Oct 01 09:23:14 crc kubenswrapper[4983]: I1001 09:23:14.060218 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerStarted","Data":"fb2f369b591c83375c0a2933cd1e3cb1009e5dae480035ad1f7564e27c89ee09"} Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.072592 4983 generic.go:334] "Generic (PLEG): container finished" podID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerID="f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9" exitCode=0 Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.072659 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerDied","Data":"f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9"} Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.929622 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9"] Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.930634 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.932796 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-bsws4" Oct 01 09:23:16 crc kubenswrapper[4983]: I1001 09:23:16.950763 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9"] Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.009410 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tbs4\" (UniqueName: \"kubernetes.io/projected/a92f120a-ff8e-479d-92ee-4ea6fc111dd0-kube-api-access-7tbs4\") pod \"rabbitmq-cluster-operator-779fc9694b-6k6c9\" (UID: \"a92f120a-ff8e-479d-92ee-4ea6fc111dd0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.081063 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerStarted","Data":"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046"} Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.099019 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b6cmc" podStartSLOduration=1.484340982 podStartE2EDuration="4.099002385s" podCreationTimestamp="2025-10-01 09:23:13 +0000 UTC" firstStartedPulling="2025-10-01 09:23:14.061184124 +0000 UTC m=+822.050412921" lastFinishedPulling="2025-10-01 09:23:16.675845527 +0000 UTC m=+824.665074324" observedRunningTime="2025-10-01 09:23:17.098637394 +0000 UTC m=+825.087866191" watchObservedRunningTime="2025-10-01 09:23:17.099002385 +0000 UTC m=+825.088231182" Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.110477 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tbs4\" (UniqueName: \"kubernetes.io/projected/a92f120a-ff8e-479d-92ee-4ea6fc111dd0-kube-api-access-7tbs4\") pod \"rabbitmq-cluster-operator-779fc9694b-6k6c9\" (UID: \"a92f120a-ff8e-479d-92ee-4ea6fc111dd0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.128062 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tbs4\" (UniqueName: \"kubernetes.io/projected/a92f120a-ff8e-479d-92ee-4ea6fc111dd0-kube-api-access-7tbs4\") pod \"rabbitmq-cluster-operator-779fc9694b-6k6c9\" (UID: \"a92f120a-ff8e-479d-92ee-4ea6fc111dd0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.244973 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" Oct 01 09:23:17 crc kubenswrapper[4983]: I1001 09:23:17.630836 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9"] Oct 01 09:23:17 crc kubenswrapper[4983]: W1001 09:23:17.639720 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda92f120a_ff8e_479d_92ee_4ea6fc111dd0.slice/crio-0ba7ad8eeebf056436a119d390b584abb4ca6c508b6ff0d4dc713ccdd641226a WatchSource:0}: Error finding container 0ba7ad8eeebf056436a119d390b584abb4ca6c508b6ff0d4dc713ccdd641226a: Status 404 returned error can't find the container with id 0ba7ad8eeebf056436a119d390b584abb4ca6c508b6ff0d4dc713ccdd641226a Oct 01 09:23:18 crc kubenswrapper[4983]: I1001 09:23:18.088065 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" event={"ID":"a92f120a-ff8e-479d-92ee-4ea6fc111dd0","Type":"ContainerStarted","Data":"0ba7ad8eeebf056436a119d390b584abb4ca6c508b6ff0d4dc713ccdd641226a"} Oct 01 09:23:20 crc kubenswrapper[4983]: I1001 09:23:20.100042 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" event={"ID":"a92f120a-ff8e-479d-92ee-4ea6fc111dd0","Type":"ContainerStarted","Data":"0b1ff00461acb2791cfece3f9474e31116c320e068b9ed8ab03d53ee69137194"} Oct 01 09:23:20 crc kubenswrapper[4983]: I1001 09:23:20.114653 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-6k6c9" podStartSLOduration=2.699077344 podStartE2EDuration="4.114635139s" podCreationTimestamp="2025-10-01 09:23:16 +0000 UTC" firstStartedPulling="2025-10-01 09:23:17.642295256 +0000 UTC m=+825.631524053" lastFinishedPulling="2025-10-01 09:23:19.057853051 +0000 UTC m=+827.047081848" observedRunningTime="2025-10-01 09:23:20.111471021 +0000 UTC m=+828.100699818" watchObservedRunningTime="2025-10-01 09:23:20.114635139 +0000 UTC m=+828.103863936" Oct 01 09:23:23 crc kubenswrapper[4983]: I1001 09:23:23.372620 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:23 crc kubenswrapper[4983]: I1001 09:23:23.373280 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:23 crc kubenswrapper[4983]: I1001 09:23:23.423165 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.178464 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.886176 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.887773 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.891635 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.891733 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.891733 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.891760 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-lkxkk" Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.892005 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 01 09:23:24 crc kubenswrapper[4983]: I1001 09:23:24.892146 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015128 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015210 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015233 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015264 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh4rm\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-kube-api-access-rh4rm\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015303 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb4c18ce-6256-49d4-8917-e50a2a87180b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015515 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb4c18ce-6256-49d4-8917-e50a2a87180b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015634 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb4c18ce-6256-49d4-8917-e50a2a87180b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.015710 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116419 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb4c18ce-6256-49d4-8917-e50a2a87180b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116471 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb4c18ce-6256-49d4-8917-e50a2a87180b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116504 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116534 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116567 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116582 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116600 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh4rm\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-kube-api-access-rh4rm\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.116619 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb4c18ce-6256-49d4-8917-e50a2a87180b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.117626 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb4c18ce-6256-49d4-8917-e50a2a87180b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.117709 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.118030 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.122418 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb4c18ce-6256-49d4-8917-e50a2a87180b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.123665 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.123693 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb4c18ce-6256-49d4-8917-e50a2a87180b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.131775 4983 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.131872 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f409543db793ed2958672d0ed5666ecb44f3c30aacd309650427fb88cd1d1290/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.135756 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh4rm\" (UniqueName: \"kubernetes.io/projected/bb4c18ce-6256-49d4-8917-e50a2a87180b-kube-api-access-rh4rm\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.162842 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a56950c0-0beb-43d4-863a-66dd6be40fc5\") pod \"rabbitmq-server-0\" (UID: \"bb4c18ce-6256-49d4-8917-e50a2a87180b\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.207039 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:23:25 crc kubenswrapper[4983]: I1001 09:23:25.609726 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 01 09:23:26 crc kubenswrapper[4983]: I1001 09:23:26.135696 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bb4c18ce-6256-49d4-8917-e50a2a87180b","Type":"ContainerStarted","Data":"9c41b228d21143ae47e84a43580c096a2c8a824f89bc5749736f064261493779"} Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.500580 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-hsdt8"] Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.501589 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.505374 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-zzbg6" Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.507015 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hsdt8"] Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.652369 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvjs6\" (UniqueName: \"kubernetes.io/projected/dc2d2684-81d6-4c72-9a28-7c553dda8c4c-kube-api-access-tvjs6\") pod \"keystone-operator-index-hsdt8\" (UID: \"dc2d2684-81d6-4c72-9a28-7c553dda8c4c\") " pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.753435 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvjs6\" (UniqueName: \"kubernetes.io/projected/dc2d2684-81d6-4c72-9a28-7c553dda8c4c-kube-api-access-tvjs6\") pod \"keystone-operator-index-hsdt8\" (UID: \"dc2d2684-81d6-4c72-9a28-7c553dda8c4c\") " pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.771197 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvjs6\" (UniqueName: \"kubernetes.io/projected/dc2d2684-81d6-4c72-9a28-7c553dda8c4c-kube-api-access-tvjs6\") pod \"keystone-operator-index-hsdt8\" (UID: \"dc2d2684-81d6-4c72-9a28-7c553dda8c4c\") " pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:27 crc kubenswrapper[4983]: I1001 09:23:27.822252 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.050190 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.050754 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b6cmc" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="registry-server" containerID="cri-o://38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046" gracePeriod=2 Oct 01 09:23:28 crc kubenswrapper[4983]: W1001 09:23:28.224156 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc2d2684_81d6_4c72_9a28_7c553dda8c4c.slice/crio-9fc59cbc462b57b1fab8d2efb6cb5557d5b0ab5349b9377a7626a60ef0bf4433 WatchSource:0}: Error finding container 9fc59cbc462b57b1fab8d2efb6cb5557d5b0ab5349b9377a7626a60ef0bf4433: Status 404 returned error can't find the container with id 9fc59cbc462b57b1fab8d2efb6cb5557d5b0ab5349b9377a7626a60ef0bf4433 Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.228651 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hsdt8"] Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.389360 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.564067 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities\") pod \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.564143 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dq7d\" (UniqueName: \"kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d\") pod \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.564262 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content\") pod \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\" (UID: \"5514ac9d-b3e8-4653-970f-bfb57c1d13c3\") " Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.566078 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities" (OuterVolumeSpecName: "utilities") pod "5514ac9d-b3e8-4653-970f-bfb57c1d13c3" (UID: "5514ac9d-b3e8-4653-970f-bfb57c1d13c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.569184 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d" (OuterVolumeSpecName: "kube-api-access-2dq7d") pod "5514ac9d-b3e8-4653-970f-bfb57c1d13c3" (UID: "5514ac9d-b3e8-4653-970f-bfb57c1d13c3"). InnerVolumeSpecName "kube-api-access-2dq7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.657949 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5514ac9d-b3e8-4653-970f-bfb57c1d13c3" (UID: "5514ac9d-b3e8-4653-970f-bfb57c1d13c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.665818 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.665848 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:28 crc kubenswrapper[4983]: I1001 09:23:28.665859 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dq7d\" (UniqueName: \"kubernetes.io/projected/5514ac9d-b3e8-4653-970f-bfb57c1d13c3-kube-api-access-2dq7d\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.159489 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hsdt8" event={"ID":"dc2d2684-81d6-4c72-9a28-7c553dda8c4c","Type":"ContainerStarted","Data":"9fc59cbc462b57b1fab8d2efb6cb5557d5b0ab5349b9377a7626a60ef0bf4433"} Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.162954 4983 generic.go:334] "Generic (PLEG): container finished" podID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerID="38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046" exitCode=0 Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.162985 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerDied","Data":"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046"} Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.163002 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6cmc" event={"ID":"5514ac9d-b3e8-4653-970f-bfb57c1d13c3","Type":"ContainerDied","Data":"fb2f369b591c83375c0a2933cd1e3cb1009e5dae480035ad1f7564e27c89ee09"} Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.163018 4983 scope.go:117] "RemoveContainer" containerID="38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046" Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.163115 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6cmc" Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.187758 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:29 crc kubenswrapper[4983]: I1001 09:23:29.196282 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b6cmc"] Oct 01 09:23:30 crc kubenswrapper[4983]: I1001 09:23:30.570482 4983 scope.go:117] "RemoveContainer" containerID="f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9" Oct 01 09:23:30 crc kubenswrapper[4983]: I1001 09:23:30.722048 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" path="/var/lib/kubelet/pods/5514ac9d-b3e8-4653-970f-bfb57c1d13c3/volumes" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.356355 4983 scope.go:117] "RemoveContainer" containerID="3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.521653 4983 scope.go:117] "RemoveContainer" containerID="38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046" Oct 01 09:23:31 crc kubenswrapper[4983]: E1001 09:23:31.522378 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046\": container with ID starting with 38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046 not found: ID does not exist" containerID="38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.522419 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046"} err="failed to get container status \"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046\": rpc error: code = NotFound desc = could not find container \"38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046\": container with ID starting with 38c74b41923ac9b857c61f70fdfe439b02db7e4279b5f08cdce3f486432bf046 not found: ID does not exist" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.522448 4983 scope.go:117] "RemoveContainer" containerID="f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9" Oct 01 09:23:31 crc kubenswrapper[4983]: E1001 09:23:31.522705 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9\": container with ID starting with f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9 not found: ID does not exist" containerID="f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.522753 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9"} err="failed to get container status \"f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9\": rpc error: code = NotFound desc = could not find container \"f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9\": container with ID starting with f66be1d1ba605712e207e3a6e932be4f3abdb941dd62b03217c029bd384de1b9 not found: ID does not exist" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.522787 4983 scope.go:117] "RemoveContainer" containerID="3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a" Oct 01 09:23:31 crc kubenswrapper[4983]: E1001 09:23:31.523083 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a\": container with ID starting with 3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a not found: ID does not exist" containerID="3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.523121 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a"} err="failed to get container status \"3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a\": rpc error: code = NotFound desc = could not find container \"3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a\": container with ID starting with 3ecb1262ee054fdacdd26f54edec3403360df96fa6efc55e0d9b941a1242fa4a not found: ID does not exist" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.961404 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.961480 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.961535 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.962250 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:23:31 crc kubenswrapper[4983]: I1001 09:23:31.962317 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7" gracePeriod=600 Oct 01 09:23:32 crc kubenswrapper[4983]: I1001 09:23:32.190293 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7" exitCode=0 Oct 01 09:23:32 crc kubenswrapper[4983]: I1001 09:23:32.190361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7"} Oct 01 09:23:32 crc kubenswrapper[4983]: I1001 09:23:32.190635 4983 scope.go:117] "RemoveContainer" containerID="d078843a9fb645cab60bcdf26c4e3ab7ffc956b7f0266d0c75d1d8be28047533" Oct 01 09:23:32 crc kubenswrapper[4983]: I1001 09:23:32.193434 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hsdt8" event={"ID":"dc2d2684-81d6-4c72-9a28-7c553dda8c4c","Type":"ContainerStarted","Data":"085f029cd3a120fbe34b5979c8adfdecace871ea16b1f71076d8ef08167055be"} Oct 01 09:23:32 crc kubenswrapper[4983]: I1001 09:23:32.212710 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-hsdt8" podStartSLOduration=1.927560083 podStartE2EDuration="5.212692093s" podCreationTimestamp="2025-10-01 09:23:27 +0000 UTC" firstStartedPulling="2025-10-01 09:23:28.229152802 +0000 UTC m=+836.218381589" lastFinishedPulling="2025-10-01 09:23:31.514284802 +0000 UTC m=+839.503513599" observedRunningTime="2025-10-01 09:23:32.208269552 +0000 UTC m=+840.197498349" watchObservedRunningTime="2025-10-01 09:23:32.212692093 +0000 UTC m=+840.201920890" Oct 01 09:23:33 crc kubenswrapper[4983]: I1001 09:23:33.205320 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025"} Oct 01 09:23:33 crc kubenswrapper[4983]: I1001 09:23:33.208003 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bb4c18ce-6256-49d4-8917-e50a2a87180b","Type":"ContainerStarted","Data":"364bb4ee0f0c06c75a4af18ae4ef137e4e92722e1e3f42ac72ce3808a29ee59c"} Oct 01 09:23:37 crc kubenswrapper[4983]: I1001 09:23:37.823202 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:37 crc kubenswrapper[4983]: I1001 09:23:37.823739 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:37 crc kubenswrapper[4983]: I1001 09:23:37.853408 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:38 crc kubenswrapper[4983]: I1001 09:23:38.264182 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-hsdt8" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.884506 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5"] Oct 01 09:23:39 crc kubenswrapper[4983]: E1001 09:23:39.884773 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="extract-content" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.884787 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="extract-content" Oct 01 09:23:39 crc kubenswrapper[4983]: E1001 09:23:39.884821 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="registry-server" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.884830 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="registry-server" Oct 01 09:23:39 crc kubenswrapper[4983]: E1001 09:23:39.884847 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="extract-utilities" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.884854 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="extract-utilities" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.884991 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5514ac9d-b3e8-4653-970f-bfb57c1d13c3" containerName="registry-server" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.886002 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.888479 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:23:39 crc kubenswrapper[4983]: I1001 09:23:39.898867 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5"] Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.026652 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.026741 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.026773 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nn9p\" (UniqueName: \"kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.127675 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.127739 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nn9p\" (UniqueName: \"kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.127790 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.128256 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.128263 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.145127 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nn9p\" (UniqueName: \"kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p\") pod \"07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.203229 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:40 crc kubenswrapper[4983]: I1001 09:23:40.667656 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5"] Oct 01 09:23:40 crc kubenswrapper[4983]: W1001 09:23:40.670384 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod155ac6db_0abb_432d_bf0d_1da60a1fab4d.slice/crio-6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3 WatchSource:0}: Error finding container 6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3: Status 404 returned error can't find the container with id 6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3 Oct 01 09:23:41 crc kubenswrapper[4983]: I1001 09:23:41.269555 4983 generic.go:334] "Generic (PLEG): container finished" podID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerID="f2056f0f542a6301177dda1da4218dbbadd57c0ccc803af398e333c7073a1575" exitCode=0 Oct 01 09:23:41 crc kubenswrapper[4983]: I1001 09:23:41.269640 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" event={"ID":"155ac6db-0abb-432d-bf0d-1da60a1fab4d","Type":"ContainerDied","Data":"f2056f0f542a6301177dda1da4218dbbadd57c0ccc803af398e333c7073a1575"} Oct 01 09:23:41 crc kubenswrapper[4983]: I1001 09:23:41.269673 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" event={"ID":"155ac6db-0abb-432d-bf0d-1da60a1fab4d","Type":"ContainerStarted","Data":"6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3"} Oct 01 09:23:44 crc kubenswrapper[4983]: I1001 09:23:44.296399 4983 generic.go:334] "Generic (PLEG): container finished" podID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerID="4a9ec8e759888f8d1a51ed74ca9d6d2ad6d36434c200245c0a892edec76637b8" exitCode=0 Oct 01 09:23:44 crc kubenswrapper[4983]: I1001 09:23:44.296483 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" event={"ID":"155ac6db-0abb-432d-bf0d-1da60a1fab4d","Type":"ContainerDied","Data":"4a9ec8e759888f8d1a51ed74ca9d6d2ad6d36434c200245c0a892edec76637b8"} Oct 01 09:23:45 crc kubenswrapper[4983]: I1001 09:23:45.304744 4983 generic.go:334] "Generic (PLEG): container finished" podID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerID="98b061670463489a01a181b9c123ce4f6ac032780a2dda2550fc27a603d542fa" exitCode=0 Oct 01 09:23:45 crc kubenswrapper[4983]: I1001 09:23:45.304863 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" event={"ID":"155ac6db-0abb-432d-bf0d-1da60a1fab4d","Type":"ContainerDied","Data":"98b061670463489a01a181b9c123ce4f6ac032780a2dda2550fc27a603d542fa"} Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.603833 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.717480 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nn9p\" (UniqueName: \"kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p\") pod \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.717879 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util\") pod \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.717919 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle\") pod \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\" (UID: \"155ac6db-0abb-432d-bf0d-1da60a1fab4d\") " Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.718755 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle" (OuterVolumeSpecName: "bundle") pod "155ac6db-0abb-432d-bf0d-1da60a1fab4d" (UID: "155ac6db-0abb-432d-bf0d-1da60a1fab4d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.722745 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p" (OuterVolumeSpecName: "kube-api-access-4nn9p") pod "155ac6db-0abb-432d-bf0d-1da60a1fab4d" (UID: "155ac6db-0abb-432d-bf0d-1da60a1fab4d"). InnerVolumeSpecName "kube-api-access-4nn9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.728951 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util" (OuterVolumeSpecName: "util") pod "155ac6db-0abb-432d-bf0d-1da60a1fab4d" (UID: "155ac6db-0abb-432d-bf0d-1da60a1fab4d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.819906 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.819945 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/155ac6db-0abb-432d-bf0d-1da60a1fab4d-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:46 crc kubenswrapper[4983]: I1001 09:23:46.819957 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nn9p\" (UniqueName: \"kubernetes.io/projected/155ac6db-0abb-432d-bf0d-1da60a1fab4d-kube-api-access-4nn9p\") on node \"crc\" DevicePath \"\"" Oct 01 09:23:47 crc kubenswrapper[4983]: I1001 09:23:47.321782 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" event={"ID":"155ac6db-0abb-432d-bf0d-1da60a1fab4d","Type":"ContainerDied","Data":"6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3"} Oct 01 09:23:47 crc kubenswrapper[4983]: I1001 09:23:47.321860 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fc2a94884f7ccd71309ea2c75ce40d817294155f3bd71db9aac9b78e38e41f3" Oct 01 09:23:47 crc kubenswrapper[4983]: I1001 09:23:47.321920 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.508479 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc"] Oct 01 09:23:57 crc kubenswrapper[4983]: E1001 09:23:57.509160 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="pull" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.509173 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="pull" Oct 01 09:23:57 crc kubenswrapper[4983]: E1001 09:23:57.509183 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="util" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.509190 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="util" Oct 01 09:23:57 crc kubenswrapper[4983]: E1001 09:23:57.509231 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="extract" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.509238 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="extract" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.509338 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="155ac6db-0abb-432d-bf0d-1da60a1fab4d" containerName="extract" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.509945 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.511881 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-v5wd6" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.512843 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.531392 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc"] Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.667770 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6xcn\" (UniqueName: \"kubernetes.io/projected/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-kube-api-access-m6xcn\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.667862 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-apiservice-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.668021 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-webhook-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.769153 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-webhook-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.769203 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6xcn\" (UniqueName: \"kubernetes.io/projected/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-kube-api-access-m6xcn\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.769231 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-apiservice-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.774181 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-webhook-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.777375 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-apiservice-cert\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.785993 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6xcn\" (UniqueName: \"kubernetes.io/projected/0f9b2567-8ee4-4aca-99f1-66971ccf2f2f-kube-api-access-m6xcn\") pod \"keystone-operator-controller-manager-f8ccbb97-ppcnc\" (UID: \"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f\") " pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:57 crc kubenswrapper[4983]: I1001 09:23:57.826269 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:23:58 crc kubenswrapper[4983]: I1001 09:23:58.063876 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc"] Oct 01 09:23:58 crc kubenswrapper[4983]: I1001 09:23:58.389890 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" event={"ID":"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f","Type":"ContainerStarted","Data":"c922de41ecdede50acfafce234825fedcb15e20608d6d221aa04b7b9ecefe157"} Oct 01 09:24:00 crc kubenswrapper[4983]: I1001 09:24:00.405667 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" event={"ID":"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f","Type":"ContainerStarted","Data":"25510cee88300b47e34bfb5731f6f0a8503d2f6a6052515b6f6c206afcc19153"} Oct 01 09:24:01 crc kubenswrapper[4983]: I1001 09:24:01.412892 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" event={"ID":"0f9b2567-8ee4-4aca-99f1-66971ccf2f2f","Type":"ContainerStarted","Data":"d59bf5016f52ffd61cc754b7215933acdba639d73cf4d701f8b32adb357b99ee"} Oct 01 09:24:01 crc kubenswrapper[4983]: I1001 09:24:01.413094 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:24:01 crc kubenswrapper[4983]: I1001 09:24:01.434760 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" podStartSLOduration=2.319139362 podStartE2EDuration="4.434736082s" podCreationTimestamp="2025-10-01 09:23:57 +0000 UTC" firstStartedPulling="2025-10-01 09:23:58.072230726 +0000 UTC m=+866.061459523" lastFinishedPulling="2025-10-01 09:24:00.187827456 +0000 UTC m=+868.177056243" observedRunningTime="2025-10-01 09:24:01.430029142 +0000 UTC m=+869.419258009" watchObservedRunningTime="2025-10-01 09:24:01.434736082 +0000 UTC m=+869.423964879" Oct 01 09:24:05 crc kubenswrapper[4983]: I1001 09:24:05.437763 4983 generic.go:334] "Generic (PLEG): container finished" podID="bb4c18ce-6256-49d4-8917-e50a2a87180b" containerID="364bb4ee0f0c06c75a4af18ae4ef137e4e92722e1e3f42ac72ce3808a29ee59c" exitCode=0 Oct 01 09:24:05 crc kubenswrapper[4983]: I1001 09:24:05.437890 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bb4c18ce-6256-49d4-8917-e50a2a87180b","Type":"ContainerDied","Data":"364bb4ee0f0c06c75a4af18ae4ef137e4e92722e1e3f42ac72ce3808a29ee59c"} Oct 01 09:24:06 crc kubenswrapper[4983]: I1001 09:24:06.447053 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bb4c18ce-6256-49d4-8917-e50a2a87180b","Type":"ContainerStarted","Data":"f96bd4b32552bcdbb023fd3b25be50a7646383f445fda570320192f57e5a6e29"} Oct 01 09:24:06 crc kubenswrapper[4983]: I1001 09:24:06.447609 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:24:06 crc kubenswrapper[4983]: I1001 09:24:06.464370 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.05916041 podStartE2EDuration="43.464347308s" podCreationTimestamp="2025-10-01 09:23:23 +0000 UTC" firstStartedPulling="2025-10-01 09:23:25.617060049 +0000 UTC m=+833.606288846" lastFinishedPulling="2025-10-01 09:23:32.022246947 +0000 UTC m=+840.011475744" observedRunningTime="2025-10-01 09:24:06.463167576 +0000 UTC m=+874.452396373" watchObservedRunningTime="2025-10-01 09:24:06.464347308 +0000 UTC m=+874.453576105" Oct 01 09:24:07 crc kubenswrapper[4983]: I1001 09:24:07.830551 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-f8ccbb97-ppcnc" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.454035 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.455087 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.457136 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-vj4hc" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.466306 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.551659 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x9v8\" (UniqueName: \"kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8\") pod \"horizon-operator-index-lnsgs\" (UID: \"2b2f9249-213f-4215-b2f7-20da5d9d9e67\") " pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.653418 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x9v8\" (UniqueName: \"kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8\") pod \"horizon-operator-index-lnsgs\" (UID: \"2b2f9249-213f-4215-b2f7-20da5d9d9e67\") " pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.672134 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x9v8\" (UniqueName: \"kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8\") pod \"horizon-operator-index-lnsgs\" (UID: \"2b2f9249-213f-4215-b2f7-20da5d9d9e67\") " pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:11 crc kubenswrapper[4983]: I1001 09:24:11.770800 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:12 crc kubenswrapper[4983]: I1001 09:24:12.178562 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:12 crc kubenswrapper[4983]: W1001 09:24:12.183662 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b2f9249_213f_4215_b2f7_20da5d9d9e67.slice/crio-95403b416c589bc75a6ee3afb93d8fe91447c8f90626895afa37d4acaf20083d WatchSource:0}: Error finding container 95403b416c589bc75a6ee3afb93d8fe91447c8f90626895afa37d4acaf20083d: Status 404 returned error can't find the container with id 95403b416c589bc75a6ee3afb93d8fe91447c8f90626895afa37d4acaf20083d Oct 01 09:24:12 crc kubenswrapper[4983]: I1001 09:24:12.486209 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-lnsgs" event={"ID":"2b2f9249-213f-4215-b2f7-20da5d9d9e67","Type":"ContainerStarted","Data":"95403b416c589bc75a6ee3afb93d8fe91447c8f90626895afa37d4acaf20083d"} Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.500668 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-lnsgs" event={"ID":"2b2f9249-213f-4215-b2f7-20da5d9d9e67","Type":"ContainerStarted","Data":"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f"} Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.517840 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-lnsgs" podStartSLOduration=2.213251073 podStartE2EDuration="3.517786332s" podCreationTimestamp="2025-10-01 09:24:11 +0000 UTC" firstStartedPulling="2025-10-01 09:24:12.186026883 +0000 UTC m=+880.175255690" lastFinishedPulling="2025-10-01 09:24:13.490562152 +0000 UTC m=+881.479790949" observedRunningTime="2025-10-01 09:24:14.515979593 +0000 UTC m=+882.505208390" watchObservedRunningTime="2025-10-01 09:24:14.517786332 +0000 UTC m=+882.507015129" Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.851308 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-sttzg"] Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.852074 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.857102 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-nl2jr" Oct 01 09:24:14 crc kubenswrapper[4983]: I1001 09:24:14.860396 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-sttzg"] Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.004934 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gmnf\" (UniqueName: \"kubernetes.io/projected/46fb7174-5de7-47bc-829c-ef6f3c0e1ce6-kube-api-access-2gmnf\") pod \"swift-operator-index-sttzg\" (UID: \"46fb7174-5de7-47bc-829c-ef6f3c0e1ce6\") " pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.106555 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gmnf\" (UniqueName: \"kubernetes.io/projected/46fb7174-5de7-47bc-829c-ef6f3c0e1ce6-kube-api-access-2gmnf\") pod \"swift-operator-index-sttzg\" (UID: \"46fb7174-5de7-47bc-829c-ef6f3c0e1ce6\") " pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.124238 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gmnf\" (UniqueName: \"kubernetes.io/projected/46fb7174-5de7-47bc-829c-ef6f3c0e1ce6-kube-api-access-2gmnf\") pod \"swift-operator-index-sttzg\" (UID: \"46fb7174-5de7-47bc-829c-ef6f3c0e1ce6\") " pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.171064 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.211540 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 01 09:24:15 crc kubenswrapper[4983]: I1001 09:24:15.588507 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-sttzg"] Oct 01 09:24:15 crc kubenswrapper[4983]: W1001 09:24:15.596874 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46fb7174_5de7_47bc_829c_ef6f3c0e1ce6.slice/crio-cd617eb8f75e7ea28f5a9259794dc051cf0fb3a08576a67868297747c3b29347 WatchSource:0}: Error finding container cd617eb8f75e7ea28f5a9259794dc051cf0fb3a08576a67868297747c3b29347: Status 404 returned error can't find the container with id cd617eb8f75e7ea28f5a9259794dc051cf0fb3a08576a67868297747c3b29347 Oct 01 09:24:16 crc kubenswrapper[4983]: I1001 09:24:16.513533 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-sttzg" event={"ID":"46fb7174-5de7-47bc-829c-ef6f3c0e1ce6","Type":"ContainerStarted","Data":"cd617eb8f75e7ea28f5a9259794dc051cf0fb3a08576a67868297747c3b29347"} Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.047442 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.047948 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-lnsgs" podUID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" containerName="registry-server" containerID="cri-o://a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f" gracePeriod=2 Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.402296 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.522894 4983 generic.go:334] "Generic (PLEG): container finished" podID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" containerID="a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f" exitCode=0 Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.522934 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-lnsgs" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.522952 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-lnsgs" event={"ID":"2b2f9249-213f-4215-b2f7-20da5d9d9e67","Type":"ContainerDied","Data":"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f"} Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.522976 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-lnsgs" event={"ID":"2b2f9249-213f-4215-b2f7-20da5d9d9e67","Type":"ContainerDied","Data":"95403b416c589bc75a6ee3afb93d8fe91447c8f90626895afa37d4acaf20083d"} Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.522993 4983 scope.go:117] "RemoveContainer" containerID="a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.524561 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-sttzg" event={"ID":"46fb7174-5de7-47bc-829c-ef6f3c0e1ce6","Type":"ContainerStarted","Data":"6a93adc2f1ef79a6d687e51ce6335ca6c9a4b188c4aa85a6c390f78629b827bb"} Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.538948 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x9v8\" (UniqueName: \"kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8\") pod \"2b2f9249-213f-4215-b2f7-20da5d9d9e67\" (UID: \"2b2f9249-213f-4215-b2f7-20da5d9d9e67\") " Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.540009 4983 scope.go:117] "RemoveContainer" containerID="a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f" Oct 01 09:24:17 crc kubenswrapper[4983]: E1001 09:24:17.540529 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f\": container with ID starting with a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f not found: ID does not exist" containerID="a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.540553 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f"} err="failed to get container status \"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f\": rpc error: code = NotFound desc = could not find container \"a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f\": container with ID starting with a39f8a5491e18de02bd9f0098a2636f312dd3042dc80ed590d0411c6c054e04f not found: ID does not exist" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.540790 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-sttzg" podStartSLOduration=2.617063217 podStartE2EDuration="3.540776107s" podCreationTimestamp="2025-10-01 09:24:14 +0000 UTC" firstStartedPulling="2025-10-01 09:24:15.59870888 +0000 UTC m=+883.587937677" lastFinishedPulling="2025-10-01 09:24:16.52242177 +0000 UTC m=+884.511650567" observedRunningTime="2025-10-01 09:24:17.536931662 +0000 UTC m=+885.526160459" watchObservedRunningTime="2025-10-01 09:24:17.540776107 +0000 UTC m=+885.530004904" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.544389 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8" (OuterVolumeSpecName: "kube-api-access-9x9v8") pod "2b2f9249-213f-4215-b2f7-20da5d9d9e67" (UID: "2b2f9249-213f-4215-b2f7-20da5d9d9e67"). InnerVolumeSpecName "kube-api-access-9x9v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.640592 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x9v8\" (UniqueName: \"kubernetes.io/projected/2b2f9249-213f-4215-b2f7-20da5d9d9e67-kube-api-access-9x9v8\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.853002 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-nvtrg"] Oct 01 09:24:17 crc kubenswrapper[4983]: E1001 09:24:17.853317 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" containerName="registry-server" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.853339 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" containerName="registry-server" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.853505 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" containerName="registry-server" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.854054 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.856436 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-vj4hc" Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.857968 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.863479 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-lnsgs"] Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.867350 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-nvtrg"] Oct 01 09:24:17 crc kubenswrapper[4983]: I1001 09:24:17.943744 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjsvh\" (UniqueName: \"kubernetes.io/projected/ed198fb7-7179-4a0a-b8f8-829cfebc0cee-kube-api-access-rjsvh\") pod \"horizon-operator-index-nvtrg\" (UID: \"ed198fb7-7179-4a0a-b8f8-829cfebc0cee\") " pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.045191 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjsvh\" (UniqueName: \"kubernetes.io/projected/ed198fb7-7179-4a0a-b8f8-829cfebc0cee-kube-api-access-rjsvh\") pod \"horizon-operator-index-nvtrg\" (UID: \"ed198fb7-7179-4a0a-b8f8-829cfebc0cee\") " pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.062028 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjsvh\" (UniqueName: \"kubernetes.io/projected/ed198fb7-7179-4a0a-b8f8-829cfebc0cee-kube-api-access-rjsvh\") pod \"horizon-operator-index-nvtrg\" (UID: \"ed198fb7-7179-4a0a-b8f8-829cfebc0cee\") " pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.170706 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.550584 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-nvtrg"] Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.722008 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b2f9249-213f-4215-b2f7-20da5d9d9e67" path="/var/lib/kubelet/pods/2b2f9249-213f-4215-b2f7-20da5d9d9e67/volumes" Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.992394 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-7tsvg"] Oct 01 09:24:18 crc kubenswrapper[4983]: I1001 09:24:18.993195 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.003835 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7tsvg"] Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.057921 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrpf\" (UniqueName: \"kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf\") pod \"keystone-db-create-7tsvg\" (UID: \"8d2667ff-ed9c-4990-866d-737077afbb35\") " pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.159008 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrpf\" (UniqueName: \"kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf\") pod \"keystone-db-create-7tsvg\" (UID: \"8d2667ff-ed9c-4990-866d-737077afbb35\") " pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.176897 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrpf\" (UniqueName: \"kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf\") pod \"keystone-db-create-7tsvg\" (UID: \"8d2667ff-ed9c-4990-866d-737077afbb35\") " pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.323943 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.543219 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-nvtrg" event={"ID":"ed198fb7-7179-4a0a-b8f8-829cfebc0cee","Type":"ContainerStarted","Data":"8f8691e99a3d80dbf099d5353545127f2679a53be21f02351bd2f26d73513c2e"} Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.543634 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-nvtrg" event={"ID":"ed198fb7-7179-4a0a-b8f8-829cfebc0cee","Type":"ContainerStarted","Data":"903c7c5282c6cf05d3657312fc8725e2066e332de97fcc37daaff0eec8baa3d9"} Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.559497 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7tsvg"] Oct 01 09:24:19 crc kubenswrapper[4983]: I1001 09:24:19.560911 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-nvtrg" podStartSLOduration=2.045976604 podStartE2EDuration="2.560897749s" podCreationTimestamp="2025-10-01 09:24:17 +0000 UTC" firstStartedPulling="2025-10-01 09:24:18.559180156 +0000 UTC m=+886.548408953" lastFinishedPulling="2025-10-01 09:24:19.074101301 +0000 UTC m=+887.063330098" observedRunningTime="2025-10-01 09:24:19.559711326 +0000 UTC m=+887.548940133" watchObservedRunningTime="2025-10-01 09:24:19.560897749 +0000 UTC m=+887.550126546" Oct 01 09:24:19 crc kubenswrapper[4983]: W1001 09:24:19.569832 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2667ff_ed9c_4990_866d_737077afbb35.slice/crio-7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022 WatchSource:0}: Error finding container 7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022: Status 404 returned error can't find the container with id 7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022 Oct 01 09:24:20 crc kubenswrapper[4983]: I1001 09:24:20.551053 4983 generic.go:334] "Generic (PLEG): container finished" podID="8d2667ff-ed9c-4990-866d-737077afbb35" containerID="470a69ffccc2a250c4241d1a7ae1c412843d0be71c005894747d77746f4ac147" exitCode=0 Oct 01 09:24:20 crc kubenswrapper[4983]: I1001 09:24:20.551123 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7tsvg" event={"ID":"8d2667ff-ed9c-4990-866d-737077afbb35","Type":"ContainerDied","Data":"470a69ffccc2a250c4241d1a7ae1c412843d0be71c005894747d77746f4ac147"} Oct 01 09:24:20 crc kubenswrapper[4983]: I1001 09:24:20.551370 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7tsvg" event={"ID":"8d2667ff-ed9c-4990-866d-737077afbb35","Type":"ContainerStarted","Data":"7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022"} Oct 01 09:24:21 crc kubenswrapper[4983]: I1001 09:24:21.824961 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:21 crc kubenswrapper[4983]: I1001 09:24:21.899149 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqrpf\" (UniqueName: \"kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf\") pod \"8d2667ff-ed9c-4990-866d-737077afbb35\" (UID: \"8d2667ff-ed9c-4990-866d-737077afbb35\") " Oct 01 09:24:21 crc kubenswrapper[4983]: I1001 09:24:21.905395 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf" (OuterVolumeSpecName: "kube-api-access-sqrpf") pod "8d2667ff-ed9c-4990-866d-737077afbb35" (UID: "8d2667ff-ed9c-4990-866d-737077afbb35"). InnerVolumeSpecName "kube-api-access-sqrpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:22 crc kubenswrapper[4983]: I1001 09:24:22.000485 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqrpf\" (UniqueName: \"kubernetes.io/projected/8d2667ff-ed9c-4990-866d-737077afbb35-kube-api-access-sqrpf\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:22 crc kubenswrapper[4983]: I1001 09:24:22.564196 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7tsvg" event={"ID":"8d2667ff-ed9c-4990-866d-737077afbb35","Type":"ContainerDied","Data":"7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022"} Oct 01 09:24:22 crc kubenswrapper[4983]: I1001 09:24:22.564243 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7195c96b7bd19555dff270589824587a4760cf5e50544a88367a65f9e89ae022" Oct 01 09:24:22 crc kubenswrapper[4983]: I1001 09:24:22.564260 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7tsvg" Oct 01 09:24:25 crc kubenswrapper[4983]: I1001 09:24:25.171936 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:25 crc kubenswrapper[4983]: I1001 09:24:25.172257 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:25 crc kubenswrapper[4983]: I1001 09:24:25.198421 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:25 crc kubenswrapper[4983]: I1001 09:24:25.611254 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-sttzg" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.171497 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.171963 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.198202 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.634355 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-nvtrg" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.891371 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-0cf5-account-create-gwzxr"] Oct 01 09:24:28 crc kubenswrapper[4983]: E1001 09:24:28.891601 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2667ff-ed9c-4990-866d-737077afbb35" containerName="mariadb-database-create" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.891612 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2667ff-ed9c-4990-866d-737077afbb35" containerName="mariadb-database-create" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.891746 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2667ff-ed9c-4990-866d-737077afbb35" containerName="mariadb-database-create" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.892156 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.894453 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.904517 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0cf5-account-create-gwzxr"] Oct 01 09:24:28 crc kubenswrapper[4983]: I1001 09:24:28.991841 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g5tx\" (UniqueName: \"kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx\") pod \"keystone-0cf5-account-create-gwzxr\" (UID: \"139a1225-84e0-4a5c-9e2f-80f7204cde88\") " pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:29 crc kubenswrapper[4983]: I1001 09:24:29.093078 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g5tx\" (UniqueName: \"kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx\") pod \"keystone-0cf5-account-create-gwzxr\" (UID: \"139a1225-84e0-4a5c-9e2f-80f7204cde88\") " pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:29 crc kubenswrapper[4983]: I1001 09:24:29.111441 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g5tx\" (UniqueName: \"kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx\") pod \"keystone-0cf5-account-create-gwzxr\" (UID: \"139a1225-84e0-4a5c-9e2f-80f7204cde88\") " pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:29 crc kubenswrapper[4983]: I1001 09:24:29.209234 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:29 crc kubenswrapper[4983]: I1001 09:24:29.619899 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0cf5-account-create-gwzxr"] Oct 01 09:24:30 crc kubenswrapper[4983]: I1001 09:24:30.616976 4983 generic.go:334] "Generic (PLEG): container finished" podID="139a1225-84e0-4a5c-9e2f-80f7204cde88" containerID="0c2677dbf7d33bf4041ffdd1108f0a23339942521d75c5ee1108b528e0adb026" exitCode=0 Oct 01 09:24:30 crc kubenswrapper[4983]: I1001 09:24:30.617026 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" event={"ID":"139a1225-84e0-4a5c-9e2f-80f7204cde88","Type":"ContainerDied","Data":"0c2677dbf7d33bf4041ffdd1108f0a23339942521d75c5ee1108b528e0adb026"} Oct 01 09:24:30 crc kubenswrapper[4983]: I1001 09:24:30.617055 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" event={"ID":"139a1225-84e0-4a5c-9e2f-80f7204cde88","Type":"ContainerStarted","Data":"0f7bf08a5a1f2990e420585fd836acdf36c1fccf50e11f37ad22b72832c15dcb"} Oct 01 09:24:31 crc kubenswrapper[4983]: I1001 09:24:31.898897 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.038671 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g5tx\" (UniqueName: \"kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx\") pod \"139a1225-84e0-4a5c-9e2f-80f7204cde88\" (UID: \"139a1225-84e0-4a5c-9e2f-80f7204cde88\") " Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.044012 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx" (OuterVolumeSpecName: "kube-api-access-7g5tx") pod "139a1225-84e0-4a5c-9e2f-80f7204cde88" (UID: "139a1225-84e0-4a5c-9e2f-80f7204cde88"). InnerVolumeSpecName "kube-api-access-7g5tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.140052 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g5tx\" (UniqueName: \"kubernetes.io/projected/139a1225-84e0-4a5c-9e2f-80f7204cde88-kube-api-access-7g5tx\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.633899 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" event={"ID":"139a1225-84e0-4a5c-9e2f-80f7204cde88","Type":"ContainerDied","Data":"0f7bf08a5a1f2990e420585fd836acdf36c1fccf50e11f37ad22b72832c15dcb"} Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.634274 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f7bf08a5a1f2990e420585fd836acdf36c1fccf50e11f37ad22b72832c15dcb" Oct 01 09:24:32 crc kubenswrapper[4983]: I1001 09:24:32.633968 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0cf5-account-create-gwzxr" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.462158 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-pqcnq"] Oct 01 09:24:34 crc kubenswrapper[4983]: E1001 09:24:34.462701 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139a1225-84e0-4a5c-9e2f-80f7204cde88" containerName="mariadb-account-create" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.462714 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="139a1225-84e0-4a5c-9e2f-80f7204cde88" containerName="mariadb-account-create" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.463461 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="139a1225-84e0-4a5c-9e2f-80f7204cde88" containerName="mariadb-account-create" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.464097 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.468278 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.468308 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.468308 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.470233 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z88s6" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.489584 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-pqcnq"] Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.572543 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.572638 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cstm7\" (UniqueName: \"kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.673406 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.673490 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cstm7\" (UniqueName: \"kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.679967 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.689509 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cstm7\" (UniqueName: \"kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7\") pod \"keystone-db-sync-pqcnq\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:34 crc kubenswrapper[4983]: I1001 09:24:34.800134 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:35 crc kubenswrapper[4983]: I1001 09:24:35.194171 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-pqcnq"] Oct 01 09:24:35 crc kubenswrapper[4983]: I1001 09:24:35.657992 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" event={"ID":"8174c1ce-cbb1-4288-a578-0478d23d3004","Type":"ContainerStarted","Data":"46c1c2ac37008f5bb599a80c882679cd4fc85e81a3a2fdaabf41944c4c4b44e1"} Oct 01 09:24:42 crc kubenswrapper[4983]: I1001 09:24:42.702144 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" event={"ID":"8174c1ce-cbb1-4288-a578-0478d23d3004","Type":"ContainerStarted","Data":"c063147fd5beee2beb6ae279fea7e356e634be0ba924530764e660e8173f9afe"} Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.705646 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" podStartSLOduration=3.877778783 podStartE2EDuration="10.705624137s" podCreationTimestamp="2025-10-01 09:24:34 +0000 UTC" firstStartedPulling="2025-10-01 09:24:35.209045908 +0000 UTC m=+903.198274715" lastFinishedPulling="2025-10-01 09:24:42.036891272 +0000 UTC m=+910.026120069" observedRunningTime="2025-10-01 09:24:42.723513116 +0000 UTC m=+910.712741913" watchObservedRunningTime="2025-10-01 09:24:44.705624137 +0000 UTC m=+912.694852944" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.708258 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg"] Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.709761 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.715780 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.725612 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg"] Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.822723 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.822769 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.822965 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwgn\" (UniqueName: \"kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.924500 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwgn\" (UniqueName: \"kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.924580 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.924611 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.925150 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.925161 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:44 crc kubenswrapper[4983]: I1001 09:24:44.950071 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwgn\" (UniqueName: \"kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn\") pod \"ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.035095 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.475740 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg"] Oct 01 09:24:45 crc kubenswrapper[4983]: W1001 09:24:45.484636 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2d46656_31cb_4e64_a979_7606de46d2fc.slice/crio-1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e WatchSource:0}: Error finding container 1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e: Status 404 returned error can't find the container with id 1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.508908 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm"] Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.511151 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.524211 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm"] Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.634580 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.634742 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjg7j\" (UniqueName: \"kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.634775 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.721883 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" event={"ID":"d2d46656-31cb-4e64-a979-7606de46d2fc","Type":"ContainerStarted","Data":"1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e"} Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.736305 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjg7j\" (UniqueName: \"kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.736347 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.736392 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.736790 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.736884 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.758919 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjg7j\" (UniqueName: \"kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j\") pod \"4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:45 crc kubenswrapper[4983]: I1001 09:24:45.927704 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.313590 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm"] Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.728272 4983 generic.go:334] "Generic (PLEG): container finished" podID="0da521dc-47ad-4443-b302-222c37f03e38" containerID="4c42d25c801ca3229502c597b7f24c51b392f05c8d7bf376f852a1f3afe77578" exitCode=0 Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.728346 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" event={"ID":"0da521dc-47ad-4443-b302-222c37f03e38","Type":"ContainerDied","Data":"4c42d25c801ca3229502c597b7f24c51b392f05c8d7bf376f852a1f3afe77578"} Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.728372 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" event={"ID":"0da521dc-47ad-4443-b302-222c37f03e38","Type":"ContainerStarted","Data":"4db3fd499ea1a237f9caa552fbb4d5e7c3faa8a0db963a0096058eeb06c97054"} Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.729894 4983 generic.go:334] "Generic (PLEG): container finished" podID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerID="2540cc03b164d8d4c751765d96a3f174c7d6579ff4a50fb98be8221cd9934b2c" exitCode=0 Oct 01 09:24:46 crc kubenswrapper[4983]: I1001 09:24:46.729988 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" event={"ID":"d2d46656-31cb-4e64-a979-7606de46d2fc","Type":"ContainerDied","Data":"2540cc03b164d8d4c751765d96a3f174c7d6579ff4a50fb98be8221cd9934b2c"} Oct 01 09:24:47 crc kubenswrapper[4983]: I1001 09:24:47.738289 4983 generic.go:334] "Generic (PLEG): container finished" podID="8174c1ce-cbb1-4288-a578-0478d23d3004" containerID="c063147fd5beee2beb6ae279fea7e356e634be0ba924530764e660e8173f9afe" exitCode=0 Oct 01 09:24:47 crc kubenswrapper[4983]: I1001 09:24:47.738375 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" event={"ID":"8174c1ce-cbb1-4288-a578-0478d23d3004","Type":"ContainerDied","Data":"c063147fd5beee2beb6ae279fea7e356e634be0ba924530764e660e8173f9afe"} Oct 01 09:24:48 crc kubenswrapper[4983]: I1001 09:24:48.753252 4983 generic.go:334] "Generic (PLEG): container finished" podID="0da521dc-47ad-4443-b302-222c37f03e38" containerID="757379dca185eb1f3f52e61e33589c6140a700616d17a406a505e31c5ffc544d" exitCode=0 Oct 01 09:24:48 crc kubenswrapper[4983]: I1001 09:24:48.753374 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" event={"ID":"0da521dc-47ad-4443-b302-222c37f03e38","Type":"ContainerDied","Data":"757379dca185eb1f3f52e61e33589c6140a700616d17a406a505e31c5ffc544d"} Oct 01 09:24:48 crc kubenswrapper[4983]: I1001 09:24:48.755476 4983 generic.go:334] "Generic (PLEG): container finished" podID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerID="65d3445105983a8fb6bdacfcb03daa1a7e2154a650439070012d00c192db43b1" exitCode=0 Oct 01 09:24:48 crc kubenswrapper[4983]: I1001 09:24:48.755538 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" event={"ID":"d2d46656-31cb-4e64-a979-7606de46d2fc","Type":"ContainerDied","Data":"65d3445105983a8fb6bdacfcb03daa1a7e2154a650439070012d00c192db43b1"} Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.169428 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.293237 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data\") pod \"8174c1ce-cbb1-4288-a578-0478d23d3004\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.293465 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cstm7\" (UniqueName: \"kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7\") pod \"8174c1ce-cbb1-4288-a578-0478d23d3004\" (UID: \"8174c1ce-cbb1-4288-a578-0478d23d3004\") " Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.300198 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7" (OuterVolumeSpecName: "kube-api-access-cstm7") pod "8174c1ce-cbb1-4288-a578-0478d23d3004" (UID: "8174c1ce-cbb1-4288-a578-0478d23d3004"). InnerVolumeSpecName "kube-api-access-cstm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.345632 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data" (OuterVolumeSpecName: "config-data") pod "8174c1ce-cbb1-4288-a578-0478d23d3004" (UID: "8174c1ce-cbb1-4288-a578-0478d23d3004"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.395368 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cstm7\" (UniqueName: \"kubernetes.io/projected/8174c1ce-cbb1-4288-a578-0478d23d3004-kube-api-access-cstm7\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.395419 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8174c1ce-cbb1-4288-a578-0478d23d3004-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.763962 4983 generic.go:334] "Generic (PLEG): container finished" podID="0da521dc-47ad-4443-b302-222c37f03e38" containerID="f95f194b4c9aa1edc64808bc7f7467ad2f82027d7ca566c4b1b771387a384d63" exitCode=0 Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.764072 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" event={"ID":"0da521dc-47ad-4443-b302-222c37f03e38","Type":"ContainerDied","Data":"f95f194b4c9aa1edc64808bc7f7467ad2f82027d7ca566c4b1b771387a384d63"} Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.767021 4983 generic.go:334] "Generic (PLEG): container finished" podID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerID="3bf69851286f84df2d5979037710b0ddab552934480af2555c416d9958f6e835" exitCode=0 Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.767091 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" event={"ID":"d2d46656-31cb-4e64-a979-7606de46d2fc","Type":"ContainerDied","Data":"3bf69851286f84df2d5979037710b0ddab552934480af2555c416d9958f6e835"} Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.768507 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" event={"ID":"8174c1ce-cbb1-4288-a578-0478d23d3004","Type":"ContainerDied","Data":"46c1c2ac37008f5bb599a80c882679cd4fc85e81a3a2fdaabf41944c4c4b44e1"} Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.768547 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-pqcnq" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.768551 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c1c2ac37008f5bb599a80c882679cd4fc85e81a3a2fdaabf41944c4c4b44e1" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.951655 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-gppdt"] Oct 01 09:24:49 crc kubenswrapper[4983]: E1001 09:24:49.951963 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8174c1ce-cbb1-4288-a578-0478d23d3004" containerName="keystone-db-sync" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.951982 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8174c1ce-cbb1-4288-a578-0478d23d3004" containerName="keystone-db-sync" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.952141 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8174c1ce-cbb1-4288-a578-0478d23d3004" containerName="keystone-db-sync" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.952648 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.957386 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.957526 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z88s6" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.957685 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.960780 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 01 09:24:49 crc kubenswrapper[4983]: I1001 09:24:49.966052 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-gppdt"] Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.003689 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh8wf\" (UniqueName: \"kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.003730 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.003762 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.003783 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.003842 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.105103 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.105147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh8wf\" (UniqueName: \"kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.105185 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.105206 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.105261 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.108574 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.108753 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.109037 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.109471 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.121443 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh8wf\" (UniqueName: \"kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf\") pod \"keystone-bootstrap-gppdt\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.266430 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.657016 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-gppdt"] Oct 01 09:24:50 crc kubenswrapper[4983]: W1001 09:24:50.666719 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad9bdb0d_abc4_4968_9906_8a25492d3369.slice/crio-14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e WatchSource:0}: Error finding container 14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e: Status 404 returned error can't find the container with id 14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e Oct 01 09:24:50 crc kubenswrapper[4983]: I1001 09:24:50.777842 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" event={"ID":"ad9bdb0d-abc4-4968-9906-8a25492d3369","Type":"ContainerStarted","Data":"14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e"} Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.091291 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.098115 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220051 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util\") pod \"d2d46656-31cb-4e64-a979-7606de46d2fc\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220137 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgwgn\" (UniqueName: \"kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn\") pod \"d2d46656-31cb-4e64-a979-7606de46d2fc\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220162 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util\") pod \"0da521dc-47ad-4443-b302-222c37f03e38\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220207 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle\") pod \"0da521dc-47ad-4443-b302-222c37f03e38\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220256 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjg7j\" (UniqueName: \"kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j\") pod \"0da521dc-47ad-4443-b302-222c37f03e38\" (UID: \"0da521dc-47ad-4443-b302-222c37f03e38\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220893 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle" (OuterVolumeSpecName: "bundle") pod "0da521dc-47ad-4443-b302-222c37f03e38" (UID: "0da521dc-47ad-4443-b302-222c37f03e38"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.220979 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle\") pod \"d2d46656-31cb-4e64-a979-7606de46d2fc\" (UID: \"d2d46656-31cb-4e64-a979-7606de46d2fc\") " Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.221660 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle" (OuterVolumeSpecName: "bundle") pod "d2d46656-31cb-4e64-a979-7606de46d2fc" (UID: "d2d46656-31cb-4e64-a979-7606de46d2fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.221919 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.221936 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.231494 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn" (OuterVolumeSpecName: "kube-api-access-zgwgn") pod "d2d46656-31cb-4e64-a979-7606de46d2fc" (UID: "d2d46656-31cb-4e64-a979-7606de46d2fc"). InnerVolumeSpecName "kube-api-access-zgwgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.232184 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j" (OuterVolumeSpecName: "kube-api-access-rjg7j") pod "0da521dc-47ad-4443-b302-222c37f03e38" (UID: "0da521dc-47ad-4443-b302-222c37f03e38"). InnerVolumeSpecName "kube-api-access-rjg7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.235160 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util" (OuterVolumeSpecName: "util") pod "0da521dc-47ad-4443-b302-222c37f03e38" (UID: "0da521dc-47ad-4443-b302-222c37f03e38"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.241773 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util" (OuterVolumeSpecName: "util") pod "d2d46656-31cb-4e64-a979-7606de46d2fc" (UID: "d2d46656-31cb-4e64-a979-7606de46d2fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.322740 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgwgn\" (UniqueName: \"kubernetes.io/projected/d2d46656-31cb-4e64-a979-7606de46d2fc-kube-api-access-zgwgn\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.322784 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0da521dc-47ad-4443-b302-222c37f03e38-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.322794 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjg7j\" (UniqueName: \"kubernetes.io/projected/0da521dc-47ad-4443-b302-222c37f03e38-kube-api-access-rjg7j\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.322802 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d46656-31cb-4e64-a979-7606de46d2fc-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.786952 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" event={"ID":"0da521dc-47ad-4443-b302-222c37f03e38","Type":"ContainerDied","Data":"4db3fd499ea1a237f9caa552fbb4d5e7c3faa8a0db963a0096058eeb06c97054"} Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.786998 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4db3fd499ea1a237f9caa552fbb4d5e7c3faa8a0db963a0096058eeb06c97054" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.787004 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.789937 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" event={"ID":"d2d46656-31cb-4e64-a979-7606de46d2fc","Type":"ContainerDied","Data":"1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e"} Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.789978 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1277c879cbc2a94f94dd0d0e0a3f4f8e374a946eeb2285e31f356a167b9f9f6e" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.789983 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg" Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.809647 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" event={"ID":"ad9bdb0d-abc4-4968-9906-8a25492d3369","Type":"ContainerStarted","Data":"b6e98d45d0505685c635bd3b875544adafd031698dd580bd2ad57c59692ea66c"} Oct 01 09:24:51 crc kubenswrapper[4983]: I1001 09:24:51.827667 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" podStartSLOduration=2.827650688 podStartE2EDuration="2.827650688s" podCreationTimestamp="2025-10-01 09:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:24:51.825637723 +0000 UTC m=+919.814866520" watchObservedRunningTime="2025-10-01 09:24:51.827650688 +0000 UTC m=+919.816879485" Oct 01 09:24:53 crc kubenswrapper[4983]: I1001 09:24:53.827337 4983 generic.go:334] "Generic (PLEG): container finished" podID="ad9bdb0d-abc4-4968-9906-8a25492d3369" containerID="b6e98d45d0505685c635bd3b875544adafd031698dd580bd2ad57c59692ea66c" exitCode=0 Oct 01 09:24:53 crc kubenswrapper[4983]: I1001 09:24:53.827406 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" event={"ID":"ad9bdb0d-abc4-4968-9906-8a25492d3369","Type":"ContainerDied","Data":"b6e98d45d0505685c635bd3b875544adafd031698dd580bd2ad57c59692ea66c"} Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.061840 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.180856 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh8wf\" (UniqueName: \"kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf\") pod \"ad9bdb0d-abc4-4968-9906-8a25492d3369\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.180910 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data\") pod \"ad9bdb0d-abc4-4968-9906-8a25492d3369\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.180929 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys\") pod \"ad9bdb0d-abc4-4968-9906-8a25492d3369\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.180963 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts\") pod \"ad9bdb0d-abc4-4968-9906-8a25492d3369\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.181049 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys\") pod \"ad9bdb0d-abc4-4968-9906-8a25492d3369\" (UID: \"ad9bdb0d-abc4-4968-9906-8a25492d3369\") " Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.185988 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ad9bdb0d-abc4-4968-9906-8a25492d3369" (UID: "ad9bdb0d-abc4-4968-9906-8a25492d3369"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.187335 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ad9bdb0d-abc4-4968-9906-8a25492d3369" (UID: "ad9bdb0d-abc4-4968-9906-8a25492d3369"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.187380 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf" (OuterVolumeSpecName: "kube-api-access-kh8wf") pod "ad9bdb0d-abc4-4968-9906-8a25492d3369" (UID: "ad9bdb0d-abc4-4968-9906-8a25492d3369"). InnerVolumeSpecName "kube-api-access-kh8wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.187386 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts" (OuterVolumeSpecName: "scripts") pod "ad9bdb0d-abc4-4968-9906-8a25492d3369" (UID: "ad9bdb0d-abc4-4968-9906-8a25492d3369"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.199975 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data" (OuterVolumeSpecName: "config-data") pod "ad9bdb0d-abc4-4968-9906-8a25492d3369" (UID: "ad9bdb0d-abc4-4968-9906-8a25492d3369"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.282791 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.282837 4983 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.282848 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.282856 4983 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad9bdb0d-abc4-4968-9906-8a25492d3369-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.282864 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh8wf\" (UniqueName: \"kubernetes.io/projected/ad9bdb0d-abc4-4968-9906-8a25492d3369-kube-api-access-kh8wf\") on node \"crc\" DevicePath \"\"" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.841534 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" event={"ID":"ad9bdb0d-abc4-4968-9906-8a25492d3369","Type":"ContainerDied","Data":"14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e"} Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.841572 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14c40225b7512b10f4bccfef94c56cdfe5a85d58e950a1af0a96e7b93489491e" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.841608 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-gppdt" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.901622 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-74bb96697f-mfssq"] Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.901938 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.901960 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.901974 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9bdb0d-abc4-4968-9906-8a25492d3369" containerName="keystone-bootstrap" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.901985 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9bdb0d-abc4-4968-9906-8a25492d3369" containerName="keystone-bootstrap" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.902006 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="util" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902015 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="util" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.902027 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="pull" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902035 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="pull" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.902048 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902055 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.902067 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="util" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902075 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="util" Oct 01 09:24:55 crc kubenswrapper[4983]: E1001 09:24:55.902113 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="pull" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902120 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="pull" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902265 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9bdb0d-abc4-4968-9906-8a25492d3369" containerName="keystone-bootstrap" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902295 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d46656-31cb-4e64-a979-7606de46d2fc" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902311 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da521dc-47ad-4443-b302-222c37f03e38" containerName="extract" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.902782 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.904918 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.905144 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.905382 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-z88s6" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.905557 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.911208 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-74bb96697f-mfssq"] Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.992096 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-scripts\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.992175 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-credential-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.992243 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mljwl\" (UniqueName: \"kubernetes.io/projected/76cb22c5-718c-4303-80ce-2e15df680917-kube-api-access-mljwl\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.992307 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-config-data\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:55 crc kubenswrapper[4983]: I1001 09:24:55.992334 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-fernet-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.093535 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-config-data\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.093589 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-fernet-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.093637 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-scripts\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.093686 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-credential-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.093739 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mljwl\" (UniqueName: \"kubernetes.io/projected/76cb22c5-718c-4303-80ce-2e15df680917-kube-api-access-mljwl\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.097087 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-credential-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.097303 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-scripts\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.097406 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-config-data\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.098249 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/76cb22c5-718c-4303-80ce-2e15df680917-fernet-keys\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.109136 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mljwl\" (UniqueName: \"kubernetes.io/projected/76cb22c5-718c-4303-80ce-2e15df680917-kube-api-access-mljwl\") pod \"keystone-74bb96697f-mfssq\" (UID: \"76cb22c5-718c-4303-80ce-2e15df680917\") " pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.223539 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.599740 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-74bb96697f-mfssq"] Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.849528 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" event={"ID":"76cb22c5-718c-4303-80ce-2e15df680917","Type":"ContainerStarted","Data":"5b4f3906d3f389a4f86c415d39a9d236f82ac08833d70ab8fbb6c5564feb4a53"} Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.849566 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" event={"ID":"76cb22c5-718c-4303-80ce-2e15df680917","Type":"ContainerStarted","Data":"1c12e7b71d49878cc65816af5cd21a509fc60716f326bfaf778cd461c67d6044"} Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.849684 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:24:56 crc kubenswrapper[4983]: I1001 09:24:56.869461 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" podStartSLOduration=1.869438129 podStartE2EDuration="1.869438129s" podCreationTimestamp="2025-10-01 09:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:24:56.864926896 +0000 UTC m=+924.854155703" watchObservedRunningTime="2025-10-01 09:24:56.869438129 +0000 UTC m=+924.858666936" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.272987 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh"] Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.274644 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.277259 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.277653 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6gd6g" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.286900 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh"] Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.316170 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx4z4\" (UniqueName: \"kubernetes.io/projected/52bd61ec-84db-4d16-8b9f-195db2a82c9c-kube-api-access-bx4z4\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.316229 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-apiservice-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.316298 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-webhook-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.417376 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx4z4\" (UniqueName: \"kubernetes.io/projected/52bd61ec-84db-4d16-8b9f-195db2a82c9c-kube-api-access-bx4z4\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.417421 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-apiservice-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.417465 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-webhook-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.422731 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-apiservice-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.424728 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/52bd61ec-84db-4d16-8b9f-195db2a82c9c-webhook-cert\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.433261 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx4z4\" (UniqueName: \"kubernetes.io/projected/52bd61ec-84db-4d16-8b9f-195db2a82c9c-kube-api-access-bx4z4\") pod \"swift-operator-controller-manager-7bfbc68795-lvtfh\" (UID: \"52bd61ec-84db-4d16-8b9f-195db2a82c9c\") " pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.592945 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:05 crc kubenswrapper[4983]: I1001 09:25:05.983504 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh"] Oct 01 09:25:06 crc kubenswrapper[4983]: I1001 09:25:06.004930 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:25:06 crc kubenswrapper[4983]: I1001 09:25:06.913165 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" event={"ID":"52bd61ec-84db-4d16-8b9f-195db2a82c9c","Type":"ContainerStarted","Data":"486a44113779a22432468eb2082c08342358d79cab2389ab81417caff4e9f53d"} Oct 01 09:25:08 crc kubenswrapper[4983]: I1001 09:25:08.927168 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" event={"ID":"52bd61ec-84db-4d16-8b9f-195db2a82c9c","Type":"ContainerStarted","Data":"02e219d896b2d59005773846cbfccda962fad5c9c17a3c45e716bb0a1e04ef80"} Oct 01 09:25:08 crc kubenswrapper[4983]: I1001 09:25:08.927496 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" event={"ID":"52bd61ec-84db-4d16-8b9f-195db2a82c9c","Type":"ContainerStarted","Data":"b69e4aa05d33b1fb91fb02e16f3bb8d5190b1d2bf7a06242853af66150ef27e0"} Oct 01 09:25:08 crc kubenswrapper[4983]: I1001 09:25:08.928532 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:08 crc kubenswrapper[4983]: I1001 09:25:08.953821 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" podStartSLOduration=1.5817405450000002 podStartE2EDuration="3.953783736s" podCreationTimestamp="2025-10-01 09:25:05 +0000 UTC" firstStartedPulling="2025-10-01 09:25:06.004635941 +0000 UTC m=+933.993864738" lastFinishedPulling="2025-10-01 09:25:08.376679132 +0000 UTC m=+936.365907929" observedRunningTime="2025-10-01 09:25:08.948571443 +0000 UTC m=+936.937800260" watchObservedRunningTime="2025-10-01 09:25:08.953783736 +0000 UTC m=+936.943012533" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.021965 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2"] Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.023559 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.030101 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.034926 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bhksl" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.036355 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2"] Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.118971 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmdh\" (UniqueName: \"kubernetes.io/projected/cda8013d-e400-4905-9533-5d0187eb33b9-kube-api-access-2bmdh\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.119038 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-apiservice-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.119094 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-webhook-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.220056 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmdh\" (UniqueName: \"kubernetes.io/projected/cda8013d-e400-4905-9533-5d0187eb33b9-kube-api-access-2bmdh\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.220099 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-apiservice-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.220171 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-webhook-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.225993 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-webhook-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.226229 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cda8013d-e400-4905-9533-5d0187eb33b9-apiservice-cert\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.237975 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmdh\" (UniqueName: \"kubernetes.io/projected/cda8013d-e400-4905-9533-5d0187eb33b9-kube-api-access-2bmdh\") pod \"horizon-operator-controller-manager-78d6c7d9cd-flwl2\" (UID: \"cda8013d-e400-4905-9533-5d0187eb33b9\") " pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.348734 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.737371 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2"] Oct 01 09:25:13 crc kubenswrapper[4983]: W1001 09:25:13.743935 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcda8013d_e400_4905_9533_5d0187eb33b9.slice/crio-2678e8c38654dd5008e6b911d922092d683b31976aa383335dc936d4f42ecf24 WatchSource:0}: Error finding container 2678e8c38654dd5008e6b911d922092d683b31976aa383335dc936d4f42ecf24: Status 404 returned error can't find the container with id 2678e8c38654dd5008e6b911d922092d683b31976aa383335dc936d4f42ecf24 Oct 01 09:25:13 crc kubenswrapper[4983]: I1001 09:25:13.959467 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" event={"ID":"cda8013d-e400-4905-9533-5d0187eb33b9","Type":"ContainerStarted","Data":"2678e8c38654dd5008e6b911d922092d683b31976aa383335dc936d4f42ecf24"} Oct 01 09:25:15 crc kubenswrapper[4983]: I1001 09:25:15.600834 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7bfbc68795-lvtfh" Oct 01 09:25:16 crc kubenswrapper[4983]: I1001 09:25:16.979276 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" event={"ID":"cda8013d-e400-4905-9533-5d0187eb33b9","Type":"ContainerStarted","Data":"8789c531838b029c2d8b9671a1a6689f8685515de24c4a57a8d847d93fd42d70"} Oct 01 09:25:16 crc kubenswrapper[4983]: I1001 09:25:16.979326 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" event={"ID":"cda8013d-e400-4905-9533-5d0187eb33b9","Type":"ContainerStarted","Data":"b33b0340cb6354a19e9a49f81505c80619cdd4403bc3e522282e9e2854363bf0"} Oct 01 09:25:16 crc kubenswrapper[4983]: I1001 09:25:16.979383 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:17 crc kubenswrapper[4983]: I1001 09:25:17.001178 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" podStartSLOduration=1.811456902 podStartE2EDuration="4.001158297s" podCreationTimestamp="2025-10-01 09:25:13 +0000 UTC" firstStartedPulling="2025-10-01 09:25:13.746199949 +0000 UTC m=+941.735428746" lastFinishedPulling="2025-10-01 09:25:15.935901354 +0000 UTC m=+943.925130141" observedRunningTime="2025-10-01 09:25:17.000380526 +0000 UTC m=+944.989609343" watchObservedRunningTime="2025-10-01 09:25:17.001158297 +0000 UTC m=+944.990387104" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.351040 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.357011 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.359231 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.359524 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.359524 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-jt98j" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.359653 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.383194 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.427975 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-lock\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.428027 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-cache\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.428068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.428086 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.428113 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66fkj\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-kube-api-access-66fkj\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530066 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530126 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530161 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66fkj\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-kube-api-access-66fkj\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530234 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-lock\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530270 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-cache\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: E1001 09:25:21.530777 4983 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:25:21 crc kubenswrapper[4983]: E1001 09:25:21.530825 4983 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:25:21 crc kubenswrapper[4983]: E1001 09:25:21.530871 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift podName:e1456183-06e2-4b5c-9ab8-c7f0986ccec6 nodeName:}" failed. No retries permitted until 2025-10-01 09:25:22.030854175 +0000 UTC m=+950.020082972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift") pod "swift-storage-0" (UID: "e1456183-06e2-4b5c-9ab8-c7f0986ccec6") : configmap "swift-ring-files" not found Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.530794 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-cache\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.531044 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-lock\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.531182 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.557856 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66fkj\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-kube-api-access-66fkj\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.570937 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.885514 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-8thz2"] Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.886582 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.889360 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.890446 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.890464 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.896372 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-8thz2"] Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935183 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935268 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935321 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935344 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935369 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:21 crc kubenswrapper[4983]: I1001 09:25:21.935472 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwxzm\" (UniqueName: \"kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.036612 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.036952 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.036978 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.037000 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwxzm\" (UniqueName: \"kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.037044 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.037082 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.037109 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:22 crc kubenswrapper[4983]: E1001 09:25:22.037204 4983 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:25:22 crc kubenswrapper[4983]: E1001 09:25:22.037216 4983 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:25:22 crc kubenswrapper[4983]: E1001 09:25:22.037254 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift podName:e1456183-06e2-4b5c-9ab8-c7f0986ccec6 nodeName:}" failed. No retries permitted until 2025-10-01 09:25:23.037240853 +0000 UTC m=+951.026469650 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift") pod "swift-storage-0" (UID: "e1456183-06e2-4b5c-9ab8-c7f0986ccec6") : configmap "swift-ring-files" not found Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.038333 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.038452 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.038890 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.042359 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.042652 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.053639 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwxzm\" (UniqueName: \"kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm\") pod \"swift-ring-rebalance-8thz2\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.215096 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:22 crc kubenswrapper[4983]: I1001 09:25:22.604312 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-8thz2"] Oct 01 09:25:22 crc kubenswrapper[4983]: W1001 09:25:22.608023 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e0c9600_5b4d_4b2d_98c4_300502282d3a.slice/crio-efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7 WatchSource:0}: Error finding container efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7: Status 404 returned error can't find the container with id efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7 Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.019798 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" event={"ID":"3e0c9600-5b4d-4b2d-98c4-300502282d3a","Type":"ContainerStarted","Data":"efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7"} Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.049395 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:23 crc kubenswrapper[4983]: E1001 09:25:23.049634 4983 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:25:23 crc kubenswrapper[4983]: E1001 09:25:23.049666 4983 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:25:23 crc kubenswrapper[4983]: E1001 09:25:23.049747 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift podName:e1456183-06e2-4b5c-9ab8-c7f0986ccec6 nodeName:}" failed. No retries permitted until 2025-10-01 09:25:25.04971812 +0000 UTC m=+953.038946937 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift") pod "swift-storage-0" (UID: "e1456183-06e2-4b5c-9ab8-c7f0986ccec6") : configmap "swift-ring-files" not found Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.064134 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-8xnbm"] Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.065330 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.067707 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-h2w64" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.077040 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-8xnbm"] Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.150793 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb48r\" (UniqueName: \"kubernetes.io/projected/aac91626-89d0-42e2-bb69-41250a58f504-kube-api-access-mb48r\") pod \"glance-operator-index-8xnbm\" (UID: \"aac91626-89d0-42e2-bb69-41250a58f504\") " pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.252822 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb48r\" (UniqueName: \"kubernetes.io/projected/aac91626-89d0-42e2-bb69-41250a58f504-kube-api-access-mb48r\") pod \"glance-operator-index-8xnbm\" (UID: \"aac91626-89d0-42e2-bb69-41250a58f504\") " pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.270931 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb48r\" (UniqueName: \"kubernetes.io/projected/aac91626-89d0-42e2-bb69-41250a58f504-kube-api-access-mb48r\") pod \"glance-operator-index-8xnbm\" (UID: \"aac91626-89d0-42e2-bb69-41250a58f504\") " pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.353135 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-78d6c7d9cd-flwl2" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.392055 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:23 crc kubenswrapper[4983]: I1001 09:25:23.790863 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-8xnbm"] Oct 01 09:25:24 crc kubenswrapper[4983]: I1001 09:25:24.026408 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-8xnbm" event={"ID":"aac91626-89d0-42e2-bb69-41250a58f504","Type":"ContainerStarted","Data":"2b505def94cdc328bb72036e0ca981493d7fec0f518f9cc29dfca6075406cf6a"} Oct 01 09:25:25 crc kubenswrapper[4983]: I1001 09:25:25.074992 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:25 crc kubenswrapper[4983]: E1001 09:25:25.075168 4983 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:25:25 crc kubenswrapper[4983]: E1001 09:25:25.075187 4983 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:25:25 crc kubenswrapper[4983]: E1001 09:25:25.075247 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift podName:e1456183-06e2-4b5c-9ab8-c7f0986ccec6 nodeName:}" failed. No retries permitted until 2025-10-01 09:25:29.075229071 +0000 UTC m=+957.064457868 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift") pod "swift-storage-0" (UID: "e1456183-06e2-4b5c-9ab8-c7f0986ccec6") : configmap "swift-ring-files" not found Oct 01 09:25:26 crc kubenswrapper[4983]: I1001 09:25:26.040407 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" event={"ID":"3e0c9600-5b4d-4b2d-98c4-300502282d3a","Type":"ContainerStarted","Data":"01cdf8b36820a1562f6d27122acb467de2f329d5c57710f07eef2f93d32411b4"} Oct 01 09:25:26 crc kubenswrapper[4983]: I1001 09:25:26.060730 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" podStartSLOduration=2.030491935 podStartE2EDuration="5.060715009s" podCreationTimestamp="2025-10-01 09:25:21 +0000 UTC" firstStartedPulling="2025-10-01 09:25:22.610128602 +0000 UTC m=+950.599357399" lastFinishedPulling="2025-10-01 09:25:25.640351666 +0000 UTC m=+953.629580473" observedRunningTime="2025-10-01 09:25:26.05894797 +0000 UTC m=+954.048176787" watchObservedRunningTime="2025-10-01 09:25:26.060715009 +0000 UTC m=+954.049943806" Oct 01 09:25:27 crc kubenswrapper[4983]: I1001 09:25:27.047405 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-8xnbm" event={"ID":"aac91626-89d0-42e2-bb69-41250a58f504","Type":"ContainerStarted","Data":"df8be18515982b94639f511d03c30dec23feee4baeba39f978d96185ddb3fdac"} Oct 01 09:25:27 crc kubenswrapper[4983]: I1001 09:25:27.068076 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-8xnbm" podStartSLOduration=1.659113963 podStartE2EDuration="4.068055664s" podCreationTimestamp="2025-10-01 09:25:23 +0000 UTC" firstStartedPulling="2025-10-01 09:25:23.802016912 +0000 UTC m=+951.791245709" lastFinishedPulling="2025-10-01 09:25:26.210958613 +0000 UTC m=+954.200187410" observedRunningTime="2025-10-01 09:25:27.06312687 +0000 UTC m=+955.052355667" watchObservedRunningTime="2025-10-01 09:25:27.068055664 +0000 UTC m=+955.057284461" Oct 01 09:25:27 crc kubenswrapper[4983]: I1001 09:25:27.661503 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-74bb96697f-mfssq" Oct 01 09:25:29 crc kubenswrapper[4983]: I1001 09:25:29.146400 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:29 crc kubenswrapper[4983]: E1001 09:25:29.146611 4983 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:25:29 crc kubenswrapper[4983]: E1001 09:25:29.146644 4983 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:25:29 crc kubenswrapper[4983]: E1001 09:25:29.146702 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift podName:e1456183-06e2-4b5c-9ab8-c7f0986ccec6 nodeName:}" failed. No retries permitted until 2025-10-01 09:25:37.146685179 +0000 UTC m=+965.135913976 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift") pod "swift-storage-0" (UID: "e1456183-06e2-4b5c-9ab8-c7f0986ccec6") : configmap "swift-ring-files" not found Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.414535 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-7578798499-ms7kc"] Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.416191 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.424624 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-7578798499-ms7kc"] Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.493903 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-log-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.493952 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-etc-swift\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.493970 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-run-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.494076 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f11862-e033-4205-90d6-44096017084e-config-data\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.494135 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8w5m\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-kube-api-access-l8w5m\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.595825 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8w5m\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-kube-api-access-l8w5m\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.595879 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-log-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.595895 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-etc-swift\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.595914 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-run-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.595987 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f11862-e033-4205-90d6-44096017084e-config-data\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.596421 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-log-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.596600 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f11862-e033-4205-90d6-44096017084e-run-httpd\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.602118 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-etc-swift\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.603567 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f11862-e033-4205-90d6-44096017084e-config-data\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.612082 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8w5m\" (UniqueName: \"kubernetes.io/projected/c7f11862-e033-4205-90d6-44096017084e-kube-api-access-l8w5m\") pod \"swift-proxy-7578798499-ms7kc\" (UID: \"c7f11862-e033-4205-90d6-44096017084e\") " pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:32 crc kubenswrapper[4983]: I1001 09:25:32.733517 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.083583 4983 generic.go:334] "Generic (PLEG): container finished" podID="3e0c9600-5b4d-4b2d-98c4-300502282d3a" containerID="01cdf8b36820a1562f6d27122acb467de2f329d5c57710f07eef2f93d32411b4" exitCode=0 Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.083870 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" event={"ID":"3e0c9600-5b4d-4b2d-98c4-300502282d3a","Type":"ContainerDied","Data":"01cdf8b36820a1562f6d27122acb467de2f329d5c57710f07eef2f93d32411b4"} Oct 01 09:25:33 crc kubenswrapper[4983]: W1001 09:25:33.166691 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7f11862_e033_4205_90d6_44096017084e.slice/crio-8270c29b61b1a698d3142003403fa3a96b764f487754eff257a4b5242594b255 WatchSource:0}: Error finding container 8270c29b61b1a698d3142003403fa3a96b764f487754eff257a4b5242594b255: Status 404 returned error can't find the container with id 8270c29b61b1a698d3142003403fa3a96b764f487754eff257a4b5242594b255 Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.169100 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-7578798499-ms7kc"] Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.392902 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.393553 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:33 crc kubenswrapper[4983]: I1001 09:25:33.436730 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.096004 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" event={"ID":"c7f11862-e033-4205-90d6-44096017084e","Type":"ContainerStarted","Data":"893364c13193000877113fe772286a9f3e8e4e34575ed486359641808e4ffc80"} Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.096429 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" event={"ID":"c7f11862-e033-4205-90d6-44096017084e","Type":"ContainerStarted","Data":"8c35bac937c7825699854664748c61d2462f333f7e0a794b7fb81ee2cd6d3f23"} Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.096453 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" event={"ID":"c7f11862-e033-4205-90d6-44096017084e","Type":"ContainerStarted","Data":"8270c29b61b1a698d3142003403fa3a96b764f487754eff257a4b5242594b255"} Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.097271 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.097322 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.113296 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" podStartSLOduration=2.113278852 podStartE2EDuration="2.113278852s" podCreationTimestamp="2025-10-01 09:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:25:34.112604353 +0000 UTC m=+962.101833150" watchObservedRunningTime="2025-10-01 09:25:34.113278852 +0000 UTC m=+962.102507649" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.130001 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-8xnbm" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.366177 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.418666 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.418778 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.418806 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.419458 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.419559 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwxzm\" (UniqueName: \"kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.419628 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices\") pod \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\" (UID: \"3e0c9600-5b4d-4b2d-98c4-300502282d3a\") " Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.420659 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.424529 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.431182 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm" (OuterVolumeSpecName: "kube-api-access-jwxzm") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "kube-api-access-jwxzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.434643 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.519754 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts" (OuterVolumeSpecName: "scripts") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.523789 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.523845 4983 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.523859 4983 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e0c9600-5b4d-4b2d-98c4-300502282d3a-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.523872 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwxzm\" (UniqueName: \"kubernetes.io/projected/3e0c9600-5b4d-4b2d-98c4-300502282d3a-kube-api-access-jwxzm\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.523884 4983 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e0c9600-5b4d-4b2d-98c4-300502282d3a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.527002 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3e0c9600-5b4d-4b2d-98c4-300502282d3a" (UID: "3e0c9600-5b4d-4b2d-98c4-300502282d3a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:25:34 crc kubenswrapper[4983]: I1001 09:25:34.625623 4983 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e0c9600-5b4d-4b2d-98c4-300502282d3a-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:35 crc kubenswrapper[4983]: I1001 09:25:35.102700 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" event={"ID":"3e0c9600-5b4d-4b2d-98c4-300502282d3a","Type":"ContainerDied","Data":"efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7"} Oct 01 09:25:35 crc kubenswrapper[4983]: I1001 09:25:35.103033 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efbb7f0033e177bd0fa3f37982c6a1eed3e111fc37cde49d9610421c0ccc27f7" Oct 01 09:25:35 crc kubenswrapper[4983]: I1001 09:25:35.102927 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-8thz2" Oct 01 09:25:37 crc kubenswrapper[4983]: I1001 09:25:37.159317 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:37 crc kubenswrapper[4983]: I1001 09:25:37.169738 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1456183-06e2-4b5c-9ab8-c7f0986ccec6-etc-swift\") pod \"swift-storage-0\" (UID: \"e1456183-06e2-4b5c-9ab8-c7f0986ccec6\") " pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:37 crc kubenswrapper[4983]: I1001 09:25:37.282477 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 01 09:25:37 crc kubenswrapper[4983]: I1001 09:25:37.705309 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 01 09:25:37 crc kubenswrapper[4983]: W1001 09:25:37.709518 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1456183_06e2_4b5c_9ab8_c7f0986ccec6.slice/crio-c78c2c309c9859c01c3bb2c2ee4f725cd476816ac13187572d6f3306372b475b WatchSource:0}: Error finding container c78c2c309c9859c01c3bb2c2ee4f725cd476816ac13187572d6f3306372b475b: Status 404 returned error can't find the container with id c78c2c309c9859c01c3bb2c2ee4f725cd476816ac13187572d6f3306372b475b Oct 01 09:25:38 crc kubenswrapper[4983]: I1001 09:25:38.122268 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"c78c2c309c9859c01c3bb2c2ee4f725cd476816ac13187572d6f3306372b475b"} Oct 01 09:25:40 crc kubenswrapper[4983]: I1001 09:25:40.136640 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"ced062fdf4c43218396a471394bd2c49c1d48e1880236dd5f2e312a6b6a8c6c8"} Oct 01 09:25:40 crc kubenswrapper[4983]: I1001 09:25:40.137101 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"0ece460f9e9b45b4e3632b4acd9f5d8455746258d8a285030d4af31288fdd404"} Oct 01 09:25:40 crc kubenswrapper[4983]: I1001 09:25:40.137114 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"41a8f9c9e8dd23c475d1dee874de8bd2545986fd3794d6f0f30db579369f5d3a"} Oct 01 09:25:40 crc kubenswrapper[4983]: I1001 09:25:40.137123 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"849e249d66975b0b7a216c06dbbbfe2dc15bc33a3fe9013cebda9965d98c3082"} Oct 01 09:25:41 crc kubenswrapper[4983]: I1001 09:25:41.149063 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"142dbc3aba26aad39cf78912356f045e1a315f7f175a371e07986de4a6209ff3"} Oct 01 09:25:41 crc kubenswrapper[4983]: I1001 09:25:41.149361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"370a6b62886e9f2bba7ec37ba86b370916c10b364ccaffff8c49c82200b6bf2e"} Oct 01 09:25:42 crc kubenswrapper[4983]: I1001 09:25:42.164191 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"bbbb2d2253fe2359a545adc083cd79d062390b9caad719ccba8e43d46b6f3d06"} Oct 01 09:25:42 crc kubenswrapper[4983]: I1001 09:25:42.164235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"bd79bbc8d6e2ada5992cc36dac6d5153f29135cb7e3035245868a53e6af0150a"} Oct 01 09:25:42 crc kubenswrapper[4983]: I1001 09:25:42.736847 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:42 crc kubenswrapper[4983]: I1001 09:25:42.737794 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-7578798499-ms7kc" Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176721 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"671fe3018dc17d55a59619a4c23fe7a613e8897d6d3a2df71b371024caf1cd40"} Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176774 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"8432a1f66114e162ad0e29fc88b4b3171cbe1b7f19fa40ca380810e29c695724"} Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176785 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"379abb6d6bfc43c416aebae8c21f0ab5ecf66b35f9f65d2e0602e1930f25f14f"} Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176795 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"01bafe62756dcd2ffb4a5eda5aa3cafd12dd492e0f800c0373404a5576cc5493"} Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176819 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"2e2628da3cc3d3cd7cd3d6e6989888c1c76fd3d4f1a74592923914f9e2a2273f"} Oct 01 09:25:43 crc kubenswrapper[4983]: I1001 09:25:43.176829 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"aa8f858ce0ae80a602e3b00ad8e292ee40d01117d7b2dba383499ed64b442a42"} Oct 01 09:25:44 crc kubenswrapper[4983]: I1001 09:25:44.189762 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"e1456183-06e2-4b5c-9ab8-c7f0986ccec6","Type":"ContainerStarted","Data":"cf5462f25bd212f6f0ec800fde3b8345c8970edefc37dff76eeb46400ef041f0"} Oct 01 09:25:44 crc kubenswrapper[4983]: I1001 09:25:44.248723 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.597653374 podStartE2EDuration="24.248697946s" podCreationTimestamp="2025-10-01 09:25:20 +0000 UTC" firstStartedPulling="2025-10-01 09:25:37.711526972 +0000 UTC m=+965.700755779" lastFinishedPulling="2025-10-01 09:25:42.362571554 +0000 UTC m=+970.351800351" observedRunningTime="2025-10-01 09:25:44.24044693 +0000 UTC m=+972.229675737" watchObservedRunningTime="2025-10-01 09:25:44.248697946 +0000 UTC m=+972.237926743" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.294250 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s"] Oct 01 09:25:47 crc kubenswrapper[4983]: E1001 09:25:47.294870 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e0c9600-5b4d-4b2d-98c4-300502282d3a" containerName="swift-ring-rebalance" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.294882 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e0c9600-5b4d-4b2d-98c4-300502282d3a" containerName="swift-ring-rebalance" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.295002 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e0c9600-5b4d-4b2d-98c4-300502282d3a" containerName="swift-ring-rebalance" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.295876 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.298125 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mlb42" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.301042 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.301122 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.301171 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkzp7\" (UniqueName: \"kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.304016 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s"] Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.401826 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkzp7\" (UniqueName: \"kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.401922 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.401981 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.402409 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.402450 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.420397 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkzp7\" (UniqueName: \"kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7\") pod \"4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:47 crc kubenswrapper[4983]: I1001 09:25:47.613210 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:48 crc kubenswrapper[4983]: I1001 09:25:48.030880 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s"] Oct 01 09:25:48 crc kubenswrapper[4983]: W1001 09:25:48.035094 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc46ba0ed_905d_4f0e_a586_30c28bbdb43a.slice/crio-e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500 WatchSource:0}: Error finding container e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500: Status 404 returned error can't find the container with id e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500 Oct 01 09:25:48 crc kubenswrapper[4983]: I1001 09:25:48.226689 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerStarted","Data":"4e42676d012e58b8e405ba4dbc5dcd0c400632daaa2b317d9c134f3e0bee0613"} Oct 01 09:25:48 crc kubenswrapper[4983]: I1001 09:25:48.227190 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerStarted","Data":"e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500"} Oct 01 09:25:49 crc kubenswrapper[4983]: I1001 09:25:49.236150 4983 generic.go:334] "Generic (PLEG): container finished" podID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerID="4e42676d012e58b8e405ba4dbc5dcd0c400632daaa2b317d9c134f3e0bee0613" exitCode=0 Oct 01 09:25:49 crc kubenswrapper[4983]: I1001 09:25:49.236221 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerDied","Data":"4e42676d012e58b8e405ba4dbc5dcd0c400632daaa2b317d9c134f3e0bee0613"} Oct 01 09:25:50 crc kubenswrapper[4983]: I1001 09:25:50.247306 4983 generic.go:334] "Generic (PLEG): container finished" podID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerID="e4bf4f93df337864fd001010d6d9ee640bef6bc9651923332ce9fa02d0a33930" exitCode=0 Oct 01 09:25:50 crc kubenswrapper[4983]: I1001 09:25:50.247379 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerDied","Data":"e4bf4f93df337864fd001010d6d9ee640bef6bc9651923332ce9fa02d0a33930"} Oct 01 09:25:51 crc kubenswrapper[4983]: I1001 09:25:51.256296 4983 generic.go:334] "Generic (PLEG): container finished" podID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerID="a2120f4eafa01768323e3c005728e2eb7af2ccdb4d6c568364f5bf52c2a8edf4" exitCode=0 Oct 01 09:25:51 crc kubenswrapper[4983]: I1001 09:25:51.256339 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerDied","Data":"a2120f4eafa01768323e3c005728e2eb7af2ccdb4d6c568364f5bf52c2a8edf4"} Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.618323 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.771395 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle\") pod \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.771941 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkzp7\" (UniqueName: \"kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7\") pod \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.772033 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle" (OuterVolumeSpecName: "bundle") pod "c46ba0ed-905d-4f0e-a586-30c28bbdb43a" (UID: "c46ba0ed-905d-4f0e-a586-30c28bbdb43a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.772198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util\") pod \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\" (UID: \"c46ba0ed-905d-4f0e-a586-30c28bbdb43a\") " Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.772767 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.777104 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7" (OuterVolumeSpecName: "kube-api-access-pkzp7") pod "c46ba0ed-905d-4f0e-a586-30c28bbdb43a" (UID: "c46ba0ed-905d-4f0e-a586-30c28bbdb43a"). InnerVolumeSpecName "kube-api-access-pkzp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.784725 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util" (OuterVolumeSpecName: "util") pod "c46ba0ed-905d-4f0e-a586-30c28bbdb43a" (UID: "c46ba0ed-905d-4f0e-a586-30c28bbdb43a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.873659 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:52 crc kubenswrapper[4983]: I1001 09:25:52.874044 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkzp7\" (UniqueName: \"kubernetes.io/projected/c46ba0ed-905d-4f0e-a586-30c28bbdb43a-kube-api-access-pkzp7\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:53 crc kubenswrapper[4983]: I1001 09:25:53.272255 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" event={"ID":"c46ba0ed-905d-4f0e-a586-30c28bbdb43a","Type":"ContainerDied","Data":"e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500"} Oct 01 09:25:53 crc kubenswrapper[4983]: I1001 09:25:53.272380 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e38bfb08e8bfb84926dd38ed8e1cfbbde4522e508916d53ae53dab6418520500" Oct 01 09:25:53 crc kubenswrapper[4983]: I1001 09:25:53.272312 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s" Oct 01 09:26:01 crc kubenswrapper[4983]: I1001 09:26:01.960762 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:26:01 crc kubenswrapper[4983]: I1001 09:26:01.961315 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.005555 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj"] Oct 01 09:26:05 crc kubenswrapper[4983]: E1001 09:26:05.006383 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="util" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.006397 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="util" Oct 01 09:26:05 crc kubenswrapper[4983]: E1001 09:26:05.006411 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="extract" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.006420 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="extract" Oct 01 09:26:05 crc kubenswrapper[4983]: E1001 09:26:05.006433 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="pull" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.006440 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="pull" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.006594 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46ba0ed-905d-4f0e-a586-30c28bbdb43a" containerName="extract" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.007445 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.009868 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gkcpx" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.010352 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.025122 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj"] Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.139408 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-webhook-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.139531 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhqr4\" (UniqueName: \"kubernetes.io/projected/c32f9b5c-913b-4cea-8850-3c71e4ae1942-kube-api-access-zhqr4\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.139657 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-apiservice-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.240500 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-webhook-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.240574 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhqr4\" (UniqueName: \"kubernetes.io/projected/c32f9b5c-913b-4cea-8850-3c71e4ae1942-kube-api-access-zhqr4\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.240633 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-apiservice-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.258316 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-webhook-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.258375 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c32f9b5c-913b-4cea-8850-3c71e4ae1942-apiservice-cert\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.260463 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhqr4\" (UniqueName: \"kubernetes.io/projected/c32f9b5c-913b-4cea-8850-3c71e4ae1942-kube-api-access-zhqr4\") pod \"glance-operator-controller-manager-bd6948d9-h6ndj\" (UID: \"c32f9b5c-913b-4cea-8850-3c71e4ae1942\") " pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.326257 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:05 crc kubenswrapper[4983]: I1001 09:26:05.740120 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj"] Oct 01 09:26:05 crc kubenswrapper[4983]: W1001 09:26:05.745351 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32f9b5c_913b_4cea_8850_3c71e4ae1942.slice/crio-ead565198ae6c315567c0bc2515a4663b362d68f87996fd3de2367cf61f328a7 WatchSource:0}: Error finding container ead565198ae6c315567c0bc2515a4663b362d68f87996fd3de2367cf61f328a7: Status 404 returned error can't find the container with id ead565198ae6c315567c0bc2515a4663b362d68f87996fd3de2367cf61f328a7 Oct 01 09:26:06 crc kubenswrapper[4983]: I1001 09:26:06.365283 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" event={"ID":"c32f9b5c-913b-4cea-8850-3c71e4ae1942","Type":"ContainerStarted","Data":"ead565198ae6c315567c0bc2515a4663b362d68f87996fd3de2367cf61f328a7"} Oct 01 09:26:07 crc kubenswrapper[4983]: I1001 09:26:07.373771 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" event={"ID":"c32f9b5c-913b-4cea-8850-3c71e4ae1942","Type":"ContainerStarted","Data":"4b55454c4e5ed606f75e7a2a029ff3a59c347433d9ce598f0eef4e31f65fa2c9"} Oct 01 09:26:08 crc kubenswrapper[4983]: I1001 09:26:08.381159 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" event={"ID":"c32f9b5c-913b-4cea-8850-3c71e4ae1942","Type":"ContainerStarted","Data":"fbefd8c899cb72695666a6beb5bc098173b8c1222ce5088eecd09525a5eabbf4"} Oct 01 09:26:08 crc kubenswrapper[4983]: I1001 09:26:08.381367 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:08 crc kubenswrapper[4983]: I1001 09:26:08.402526 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" podStartSLOduration=2.682915927 podStartE2EDuration="4.402509459s" podCreationTimestamp="2025-10-01 09:26:04 +0000 UTC" firstStartedPulling="2025-10-01 09:26:05.747530771 +0000 UTC m=+993.736759568" lastFinishedPulling="2025-10-01 09:26:07.467124303 +0000 UTC m=+995.456353100" observedRunningTime="2025-10-01 09:26:08.398482679 +0000 UTC m=+996.387711476" watchObservedRunningTime="2025-10-01 09:26:08.402509459 +0000 UTC m=+996.391738256" Oct 01 09:26:15 crc kubenswrapper[4983]: I1001 09:26:15.330616 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-bd6948d9-h6ndj" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.509755 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.510947 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.513499 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.513505 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.514108 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-c44s5" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.514573 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.519605 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.548248 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-r5bzq"] Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.549065 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.556096 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-r5bzq"] Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.642112 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.642193 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.642543 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.642600 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ltzj\" (UniqueName: \"kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.743908 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqpv7\" (UniqueName: \"kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7\") pod \"glance-db-create-r5bzq\" (UID: \"a4091bd7-099e-4663-abdd-189b8131920b\") " pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.743977 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.744054 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.744092 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ltzj\" (UniqueName: \"kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.744196 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.745240 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.746075 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.761671 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.771087 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ltzj\" (UniqueName: \"kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj\") pod \"openstackclient\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.827554 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.845334 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqpv7\" (UniqueName: \"kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7\") pod \"glance-db-create-r5bzq\" (UID: \"a4091bd7-099e-4663-abdd-189b8131920b\") " pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:20 crc kubenswrapper[4983]: I1001 09:26:20.862718 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqpv7\" (UniqueName: \"kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7\") pod \"glance-db-create-r5bzq\" (UID: \"a4091bd7-099e-4663-abdd-189b8131920b\") " pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:21 crc kubenswrapper[4983]: I1001 09:26:21.102820 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:26:21 crc kubenswrapper[4983]: W1001 09:26:21.115905 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod593a0d0b_be96_4c44_8f7e_6741b06554ee.slice/crio-5b69bffcb48a0e3f9517170623e1e1f786637979432b0b8b317493ed5fd7d70e WatchSource:0}: Error finding container 5b69bffcb48a0e3f9517170623e1e1f786637979432b0b8b317493ed5fd7d70e: Status 404 returned error can't find the container with id 5b69bffcb48a0e3f9517170623e1e1f786637979432b0b8b317493ed5fd7d70e Oct 01 09:26:21 crc kubenswrapper[4983]: I1001 09:26:21.162241 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:21 crc kubenswrapper[4983]: I1001 09:26:21.472728 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"593a0d0b-be96-4c44-8f7e-6741b06554ee","Type":"ContainerStarted","Data":"5b69bffcb48a0e3f9517170623e1e1f786637979432b0b8b317493ed5fd7d70e"} Oct 01 09:26:21 crc kubenswrapper[4983]: I1001 09:26:21.567512 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-r5bzq"] Oct 01 09:26:21 crc kubenswrapper[4983]: W1001 09:26:21.570210 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4091bd7_099e_4663_abdd_189b8131920b.slice/crio-1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf WatchSource:0}: Error finding container 1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf: Status 404 returned error can't find the container with id 1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf Oct 01 09:26:22 crc kubenswrapper[4983]: I1001 09:26:22.491637 4983 generic.go:334] "Generic (PLEG): container finished" podID="a4091bd7-099e-4663-abdd-189b8131920b" containerID="be88385fa74c8a6e25ecd1d2ac544f0bd5e40dbe2c2d2541cc0080f0bef6f017" exitCode=0 Oct 01 09:26:22 crc kubenswrapper[4983]: I1001 09:26:22.491697 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-r5bzq" event={"ID":"a4091bd7-099e-4663-abdd-189b8131920b","Type":"ContainerDied","Data":"be88385fa74c8a6e25ecd1d2ac544f0bd5e40dbe2c2d2541cc0080f0bef6f017"} Oct 01 09:26:22 crc kubenswrapper[4983]: I1001 09:26:22.491971 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-r5bzq" event={"ID":"a4091bd7-099e-4663-abdd-189b8131920b","Type":"ContainerStarted","Data":"1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf"} Oct 01 09:26:23 crc kubenswrapper[4983]: I1001 09:26:23.782839 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:23 crc kubenswrapper[4983]: I1001 09:26:23.787682 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqpv7\" (UniqueName: \"kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7\") pod \"a4091bd7-099e-4663-abdd-189b8131920b\" (UID: \"a4091bd7-099e-4663-abdd-189b8131920b\") " Oct 01 09:26:23 crc kubenswrapper[4983]: I1001 09:26:23.798983 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7" (OuterVolumeSpecName: "kube-api-access-mqpv7") pod "a4091bd7-099e-4663-abdd-189b8131920b" (UID: "a4091bd7-099e-4663-abdd-189b8131920b"). InnerVolumeSpecName "kube-api-access-mqpv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:26:23 crc kubenswrapper[4983]: I1001 09:26:23.888961 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqpv7\" (UniqueName: \"kubernetes.io/projected/a4091bd7-099e-4663-abdd-189b8131920b-kube-api-access-mqpv7\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:24 crc kubenswrapper[4983]: I1001 09:26:24.507339 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-r5bzq" event={"ID":"a4091bd7-099e-4663-abdd-189b8131920b","Type":"ContainerDied","Data":"1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf"} Oct 01 09:26:24 crc kubenswrapper[4983]: I1001 09:26:24.507672 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee8e7d4a14af95e7277e1e273a58db27f928e991b8c1bf4460a06879c18a1cf" Oct 01 09:26:24 crc kubenswrapper[4983]: I1001 09:26:24.507395 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-r5bzq" Oct 01 09:26:29 crc kubenswrapper[4983]: I1001 09:26:29.541600 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"593a0d0b-be96-4c44-8f7e-6741b06554ee","Type":"ContainerStarted","Data":"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c"} Oct 01 09:26:29 crc kubenswrapper[4983]: I1001 09:26:29.560802 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.7517319420000002 podStartE2EDuration="9.560782918s" podCreationTimestamp="2025-10-01 09:26:20 +0000 UTC" firstStartedPulling="2025-10-01 09:26:21.119317133 +0000 UTC m=+1009.108545930" lastFinishedPulling="2025-10-01 09:26:28.928368109 +0000 UTC m=+1016.917596906" observedRunningTime="2025-10-01 09:26:29.555340188 +0000 UTC m=+1017.544569015" watchObservedRunningTime="2025-10-01 09:26:29.560782918 +0000 UTC m=+1017.550011735" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.553681 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f9e7-account-create-gnh4k"] Oct 01 09:26:30 crc kubenswrapper[4983]: E1001 09:26:30.554012 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4091bd7-099e-4663-abdd-189b8131920b" containerName="mariadb-database-create" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.554025 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4091bd7-099e-4663-abdd-189b8131920b" containerName="mariadb-database-create" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.554173 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4091bd7-099e-4663-abdd-189b8131920b" containerName="mariadb-database-create" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.554631 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.558803 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.564505 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f9e7-account-create-gnh4k"] Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.681151 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2qmh\" (UniqueName: \"kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh\") pod \"glance-f9e7-account-create-gnh4k\" (UID: \"9d4c026f-de4c-4a34-af25-a41370a5af55\") " pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.782531 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2qmh\" (UniqueName: \"kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh\") pod \"glance-f9e7-account-create-gnh4k\" (UID: \"9d4c026f-de4c-4a34-af25-a41370a5af55\") " pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.799563 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2qmh\" (UniqueName: \"kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh\") pod \"glance-f9e7-account-create-gnh4k\" (UID: \"9d4c026f-de4c-4a34-af25-a41370a5af55\") " pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:30 crc kubenswrapper[4983]: I1001 09:26:30.868412 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.296396 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f9e7-account-create-gnh4k"] Oct 01 09:26:31 crc kubenswrapper[4983]: W1001 09:26:31.301937 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d4c026f_de4c_4a34_af25_a41370a5af55.slice/crio-57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58 WatchSource:0}: Error finding container 57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58: Status 404 returned error can't find the container with id 57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58 Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.556127 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" event={"ID":"9d4c026f-de4c-4a34-af25-a41370a5af55","Type":"ContainerStarted","Data":"750db2a684d3f813b2811ae3e5aeb15fdcbc0abbd62dbaf3f34669faa4561f38"} Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.556207 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" event={"ID":"9d4c026f-de4c-4a34-af25-a41370a5af55","Type":"ContainerStarted","Data":"57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58"} Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.573215 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" podStartSLOduration=1.573198119 podStartE2EDuration="1.573198119s" podCreationTimestamp="2025-10-01 09:26:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:26:31.569470977 +0000 UTC m=+1019.558699774" watchObservedRunningTime="2025-10-01 09:26:31.573198119 +0000 UTC m=+1019.562426916" Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.960739 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:26:31 crc kubenswrapper[4983]: I1001 09:26:31.960877 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:26:32 crc kubenswrapper[4983]: I1001 09:26:32.565675 4983 generic.go:334] "Generic (PLEG): container finished" podID="9d4c026f-de4c-4a34-af25-a41370a5af55" containerID="750db2a684d3f813b2811ae3e5aeb15fdcbc0abbd62dbaf3f34669faa4561f38" exitCode=0 Oct 01 09:26:32 crc kubenswrapper[4983]: I1001 09:26:32.565750 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" event={"ID":"9d4c026f-de4c-4a34-af25-a41370a5af55","Type":"ContainerDied","Data":"750db2a684d3f813b2811ae3e5aeb15fdcbc0abbd62dbaf3f34669faa4561f38"} Oct 01 09:26:33 crc kubenswrapper[4983]: I1001 09:26:33.810538 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:33 crc kubenswrapper[4983]: I1001 09:26:33.925273 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2qmh\" (UniqueName: \"kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh\") pod \"9d4c026f-de4c-4a34-af25-a41370a5af55\" (UID: \"9d4c026f-de4c-4a34-af25-a41370a5af55\") " Oct 01 09:26:33 crc kubenswrapper[4983]: I1001 09:26:33.930682 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh" (OuterVolumeSpecName: "kube-api-access-m2qmh") pod "9d4c026f-de4c-4a34-af25-a41370a5af55" (UID: "9d4c026f-de4c-4a34-af25-a41370a5af55"). InnerVolumeSpecName "kube-api-access-m2qmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:26:34 crc kubenswrapper[4983]: I1001 09:26:34.027062 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2qmh\" (UniqueName: \"kubernetes.io/projected/9d4c026f-de4c-4a34-af25-a41370a5af55-kube-api-access-m2qmh\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:34 crc kubenswrapper[4983]: I1001 09:26:34.581039 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" event={"ID":"9d4c026f-de4c-4a34-af25-a41370a5af55","Type":"ContainerDied","Data":"57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58"} Oct 01 09:26:34 crc kubenswrapper[4983]: I1001 09:26:34.581079 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57217676bff79feb71ce8fdc383c7cbc4b3594183436bca76e42507a1ffb1b58" Oct 01 09:26:34 crc kubenswrapper[4983]: I1001 09:26:34.581155 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f9e7-account-create-gnh4k" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.617700 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-848kt"] Oct 01 09:26:35 crc kubenswrapper[4983]: E1001 09:26:35.618252 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4c026f-de4c-4a34-af25-a41370a5af55" containerName="mariadb-account-create" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.618265 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4c026f-de4c-4a34-af25-a41370a5af55" containerName="mariadb-account-create" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.618405 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4c026f-de4c-4a34-af25-a41370a5af55" containerName="mariadb-account-create" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.618864 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.621079 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-8dwg7" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.621533 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.625367 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-848kt"] Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.751870 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.751961 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nbww\" (UniqueName: \"kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.752015 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.853482 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nbww\" (UniqueName: \"kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.856917 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.857737 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.862677 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.863107 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.869971 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nbww\" (UniqueName: \"kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww\") pod \"glance-db-sync-848kt\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:35 crc kubenswrapper[4983]: I1001 09:26:35.944335 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:36 crc kubenswrapper[4983]: I1001 09:26:36.335548 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-848kt"] Oct 01 09:26:36 crc kubenswrapper[4983]: I1001 09:26:36.598071 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-848kt" event={"ID":"4de3d787-7a87-4578-ad12-3e338a474309","Type":"ContainerStarted","Data":"27be6c465ad5f5e917ecc10d1ce5361d81f28991f6ebf18d642b8c0c14d1161e"} Oct 01 09:26:47 crc kubenswrapper[4983]: I1001 09:26:47.679515 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-848kt" event={"ID":"4de3d787-7a87-4578-ad12-3e338a474309","Type":"ContainerStarted","Data":"f5cf9c616c03b16b472300f1068b9e63e581b7a4115b8924061fd1137a9a77ae"} Oct 01 09:26:47 crc kubenswrapper[4983]: I1001 09:26:47.702753 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-848kt" podStartSLOduration=2.337459386 podStartE2EDuration="12.702731014s" podCreationTimestamp="2025-10-01 09:26:35 +0000 UTC" firstStartedPulling="2025-10-01 09:26:36.343100185 +0000 UTC m=+1024.332328982" lastFinishedPulling="2025-10-01 09:26:46.708371773 +0000 UTC m=+1034.697600610" observedRunningTime="2025-10-01 09:26:47.692243547 +0000 UTC m=+1035.681472384" watchObservedRunningTime="2025-10-01 09:26:47.702731014 +0000 UTC m=+1035.691959811" Oct 01 09:26:53 crc kubenswrapper[4983]: I1001 09:26:53.722315 4983 generic.go:334] "Generic (PLEG): container finished" podID="4de3d787-7a87-4578-ad12-3e338a474309" containerID="f5cf9c616c03b16b472300f1068b9e63e581b7a4115b8924061fd1137a9a77ae" exitCode=0 Oct 01 09:26:53 crc kubenswrapper[4983]: I1001 09:26:53.722416 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-848kt" event={"ID":"4de3d787-7a87-4578-ad12-3e338a474309","Type":"ContainerDied","Data":"f5cf9c616c03b16b472300f1068b9e63e581b7a4115b8924061fd1137a9a77ae"} Oct 01 09:26:54 crc kubenswrapper[4983]: I1001 09:26:54.958930 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.132517 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data\") pod \"4de3d787-7a87-4578-ad12-3e338a474309\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.132592 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nbww\" (UniqueName: \"kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww\") pod \"4de3d787-7a87-4578-ad12-3e338a474309\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.132638 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data\") pod \"4de3d787-7a87-4578-ad12-3e338a474309\" (UID: \"4de3d787-7a87-4578-ad12-3e338a474309\") " Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.137272 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww" (OuterVolumeSpecName: "kube-api-access-7nbww") pod "4de3d787-7a87-4578-ad12-3e338a474309" (UID: "4de3d787-7a87-4578-ad12-3e338a474309"). InnerVolumeSpecName "kube-api-access-7nbww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.137503 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4de3d787-7a87-4578-ad12-3e338a474309" (UID: "4de3d787-7a87-4578-ad12-3e338a474309"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.168216 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data" (OuterVolumeSpecName: "config-data") pod "4de3d787-7a87-4578-ad12-3e338a474309" (UID: "4de3d787-7a87-4578-ad12-3e338a474309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.234083 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.234118 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nbww\" (UniqueName: \"kubernetes.io/projected/4de3d787-7a87-4578-ad12-3e338a474309-kube-api-access-7nbww\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.234131 4983 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4de3d787-7a87-4578-ad12-3e338a474309-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.737274 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-848kt" event={"ID":"4de3d787-7a87-4578-ad12-3e338a474309","Type":"ContainerDied","Data":"27be6c465ad5f5e917ecc10d1ce5361d81f28991f6ebf18d642b8c0c14d1161e"} Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.737579 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27be6c465ad5f5e917ecc10d1ce5361d81f28991f6ebf18d642b8c0c14d1161e" Oct 01 09:26:55 crc kubenswrapper[4983]: I1001 09:26:55.737326 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-848kt" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.069610 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:26:57 crc kubenswrapper[4983]: E1001 09:26:57.070222 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4de3d787-7a87-4578-ad12-3e338a474309" containerName="glance-db-sync" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.070236 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="4de3d787-7a87-4578-ad12-3e338a474309" containerName="glance-db-sync" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.070377 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="4de3d787-7a87-4578-ad12-3e338a474309" containerName="glance-db-sync" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.071117 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.073559 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.074285 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.074459 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-8dwg7" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.109505 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.117608 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.118847 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.127738 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160017 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160071 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160126 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160180 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160206 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160231 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160251 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160271 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ctwm\" (UniqueName: \"kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160303 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160334 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160526 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160649 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160681 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.160719 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.261835 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.261939 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.261984 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262007 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262037 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262072 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262090 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262137 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262209 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc8df\" (UniqueName: \"kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262274 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262299 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262317 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262325 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262352 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262469 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262498 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262526 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262543 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262557 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262579 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262583 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262597 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262638 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262653 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262669 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ctwm\" (UniqueName: \"kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262698 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262707 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262724 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262753 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262767 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262825 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262861 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262890 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.262924 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.263404 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.263452 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.263540 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.263687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.263716 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.278435 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.281616 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.289492 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ctwm\" (UniqueName: \"kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.297049 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.317544 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.355428 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:57 crc kubenswrapper[4983]: E1001 09:26:57.355930 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-kc8df lib-modules logs run scripts sys var-locks-brick], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-single-1" podUID="59f8c9d2-b28c-464e-a774-c11227d140c4" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364541 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364586 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364608 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364623 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364666 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364874 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364888 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364949 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364900 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365295 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365307 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365358 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365409 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365448 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365492 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365540 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365575 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc8df\" (UniqueName: \"kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365639 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.364658 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365738 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365758 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365775 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365796 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.365830 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.369346 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.371391 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.387298 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.387445 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc8df\" (UniqueName: \"kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.391175 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.392674 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.751082 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.762062 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.821334 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873564 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc8df\" (UniqueName: \"kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873670 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873694 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873708 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873754 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873772 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873787 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873863 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873900 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873918 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873919 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873938 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873968 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.873996 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874050 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874072 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874107 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874151 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run\") pod \"59f8c9d2-b28c-464e-a774-c11227d140c4\" (UID: \"59f8c9d2-b28c-464e-a774-c11227d140c4\") " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874422 4983 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874439 4983 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874449 4983 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874458 4983 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874456 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run" (OuterVolumeSpecName: "run") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874666 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.874716 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs" (OuterVolumeSpecName: "logs") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.875065 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys" (OuterVolumeSpecName: "sys") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.875106 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev" (OuterVolumeSpecName: "dev") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.877515 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts" (OuterVolumeSpecName: "scripts") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.877525 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.878132 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df" (OuterVolumeSpecName: "kube-api-access-kc8df") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "kube-api-access-kc8df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.878709 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.879383 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data" (OuterVolumeSpecName: "config-data") pod "59f8c9d2-b28c-464e-a774-c11227d140c4" (UID: "59f8c9d2-b28c-464e-a774-c11227d140c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976043 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976080 4983 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976090 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976098 4983 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-sys\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976107 4983 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/59f8c9d2-b28c-464e-a774-c11227d140c4-dev\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976116 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976126 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc8df\" (UniqueName: \"kubernetes.io/projected/59f8c9d2-b28c-464e-a774-c11227d140c4-kube-api-access-kc8df\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976136 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f8c9d2-b28c-464e-a774-c11227d140c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976146 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f8c9d2-b28c-464e-a774-c11227d140c4-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.976167 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.988931 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 01 09:26:57 crc kubenswrapper[4983]: I1001 09:26:57.989337 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.077003 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.077039 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.761752 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.761749 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerStarted","Data":"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375"} Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.762709 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerStarted","Data":"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44"} Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.762732 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerStarted","Data":"ba362733968ccddf228f8b08824de8e548c2847b1aa8becb248e1298d760a7c5"} Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.786642 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.78661168 podStartE2EDuration="2.78661168s" podCreationTimestamp="2025-10-01 09:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:26:58.782851668 +0000 UTC m=+1046.772080485" watchObservedRunningTime="2025-10-01 09:26:58.78661168 +0000 UTC m=+1046.775840517" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.826095 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.831103 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.849962 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.851362 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.877951 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.993238 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.993469 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.993648 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.993946 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994112 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994132 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994227 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994630 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.994985 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.995100 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.995156 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.995271 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:58 crc kubenswrapper[4983]: I1001 09:26:58.995326 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xft56\" (UniqueName: \"kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097239 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097295 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097322 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097341 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097375 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097401 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xft56\" (UniqueName: \"kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097426 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097450 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097485 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097513 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097556 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097554 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097432 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097640 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097578 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097716 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097766 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097846 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097892 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097906 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.097980 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.098001 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.098168 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.098271 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.098530 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.105559 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.105759 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.116578 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xft56\" (UniqueName: \"kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.135382 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.143688 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.171271 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.602484 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:26:59 crc kubenswrapper[4983]: W1001 09:26:59.612787 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc79d04e6_6bf4_4baa_a6b3_cce69ae405ce.slice/crio-5357d795c28ed14c98d82cb4da5a20f4f31b9b9df8353a1c6cc731629ad6288f WatchSource:0}: Error finding container 5357d795c28ed14c98d82cb4da5a20f4f31b9b9df8353a1c6cc731629ad6288f: Status 404 returned error can't find the container with id 5357d795c28ed14c98d82cb4da5a20f4f31b9b9df8353a1c6cc731629ad6288f Oct 01 09:26:59 crc kubenswrapper[4983]: I1001 09:26:59.769582 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerStarted","Data":"5357d795c28ed14c98d82cb4da5a20f4f31b9b9df8353a1c6cc731629ad6288f"} Oct 01 09:27:00 crc kubenswrapper[4983]: I1001 09:27:00.722405 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f8c9d2-b28c-464e-a774-c11227d140c4" path="/var/lib/kubelet/pods/59f8c9d2-b28c-464e-a774-c11227d140c4/volumes" Oct 01 09:27:00 crc kubenswrapper[4983]: I1001 09:27:00.776762 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerStarted","Data":"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2"} Oct 01 09:27:00 crc kubenswrapper[4983]: I1001 09:27:00.776822 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerStarted","Data":"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9"} Oct 01 09:27:00 crc kubenswrapper[4983]: I1001 09:27:00.799125 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.799102714 podStartE2EDuration="2.799102714s" podCreationTimestamp="2025-10-01 09:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:27:00.793938534 +0000 UTC m=+1048.783167341" watchObservedRunningTime="2025-10-01 09:27:00.799102714 +0000 UTC m=+1048.788331511" Oct 01 09:27:01 crc kubenswrapper[4983]: I1001 09:27:01.960594 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:27:01 crc kubenswrapper[4983]: I1001 09:27:01.960659 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:27:01 crc kubenswrapper[4983]: I1001 09:27:01.960707 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:27:01 crc kubenswrapper[4983]: I1001 09:27:01.961543 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:27:01 crc kubenswrapper[4983]: I1001 09:27:01.961611 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025" gracePeriod=600 Oct 01 09:27:02 crc kubenswrapper[4983]: I1001 09:27:02.791315 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025" exitCode=0 Oct 01 09:27:02 crc kubenswrapper[4983]: I1001 09:27:02.791962 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025"} Oct 01 09:27:02 crc kubenswrapper[4983]: I1001 09:27:02.792005 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50"} Oct 01 09:27:02 crc kubenswrapper[4983]: I1001 09:27:02.792021 4983 scope.go:117] "RemoveContainer" containerID="2a49667d09e33c0bc9bbdad56c58fe3b7254b3e316cf4ef099c43dba2f109ea7" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.392066 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.392637 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.415297 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.426779 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.830951 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:07 crc kubenswrapper[4983]: I1001 09:27:07.830993 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.172231 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.172531 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.193845 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.205635 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.842672 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.842826 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.962197 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.962286 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:27:09 crc kubenswrapper[4983]: I1001 09:27:09.965876 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:11 crc kubenswrapper[4983]: I1001 09:27:11.997656 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:11 crc kubenswrapper[4983]: I1001 09:27:11.998259 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.003351 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.062752 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.063018 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-log" containerID="cri-o://89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44" gracePeriod=30 Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.063465 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-httpd" containerID="cri-o://d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375" gracePeriod=30 Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.070096 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": EOF" Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.070260 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": EOF" Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.865775 4983 generic.go:334] "Generic (PLEG): container finished" podID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerID="89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44" exitCode=143 Oct 01 09:27:12 crc kubenswrapper[4983]: I1001 09:27:12.865845 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerDied","Data":"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44"} Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.777123 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.888956 4983 generic.go:334] "Generic (PLEG): container finished" podID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerID="d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375" exitCode=0 Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.889013 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerDied","Data":"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375"} Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.889027 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.889062 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2","Type":"ContainerDied","Data":"ba362733968ccddf228f8b08824de8e548c2847b1aa8becb248e1298d760a7c5"} Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.889086 4983 scope.go:117] "RemoveContainer" containerID="d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.911215 4983 scope.go:117] "RemoveContainer" containerID="89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.929244 4983 scope.go:117] "RemoveContainer" containerID="d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375" Oct 01 09:27:15 crc kubenswrapper[4983]: E1001 09:27:15.929866 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375\": container with ID starting with d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375 not found: ID does not exist" containerID="d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.929924 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375"} err="failed to get container status \"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375\": rpc error: code = NotFound desc = could not find container \"d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375\": container with ID starting with d07fcd276d8c44c8dec0fe99b8ebb4cb0707680fe1dd34754ab4f7ff326bc375 not found: ID does not exist" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.929959 4983 scope.go:117] "RemoveContainer" containerID="89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44" Oct 01 09:27:15 crc kubenswrapper[4983]: E1001 09:27:15.930409 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44\": container with ID starting with 89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44 not found: ID does not exist" containerID="89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.930450 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44"} err="failed to get container status \"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44\": rpc error: code = NotFound desc = could not find container \"89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44\": container with ID starting with 89e760ed397858851f35b5e94e9c823561f9d51c89ffb682f23bd811f08ebb44 not found: ID does not exist" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937022 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ctwm\" (UniqueName: \"kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937090 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937135 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937168 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937267 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937307 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937351 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937388 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937414 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937452 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937500 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.937535 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\" (UID: \"f10fdfbf-dec4-4371-bc6c-a4454f7b37e2\") " Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938287 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938605 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938634 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs" (OuterVolumeSpecName: "logs") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938682 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run" (OuterVolumeSpecName: "run") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938605 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938747 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.938828 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev" (OuterVolumeSpecName: "dev") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.939041 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys" (OuterVolumeSpecName: "sys") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.939886 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.943012 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm" (OuterVolumeSpecName: "kube-api-access-7ctwm") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "kube-api-access-7ctwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.943247 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts" (OuterVolumeSpecName: "scripts") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.954558 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.956929 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:15 crc kubenswrapper[4983]: I1001 09:27:15.975540 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data" (OuterVolumeSpecName: "config-data") pod "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" (UID: "f10fdfbf-dec4-4371-bc6c-a4454f7b37e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039154 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039201 4983 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039211 4983 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039222 4983 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039254 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039269 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039277 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039284 4983 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039303 4983 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039312 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039320 4983 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-dev\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039329 4983 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-sys\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039341 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.039351 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ctwm\" (UniqueName: \"kubernetes.io/projected/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2-kube-api-access-7ctwm\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.053455 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.055609 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.140945 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.140984 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.216895 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.220844 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.234904 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:16 crc kubenswrapper[4983]: E1001 09:27:16.235195 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-httpd" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.235210 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-httpd" Oct 01 09:27:16 crc kubenswrapper[4983]: E1001 09:27:16.235223 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-log" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.235229 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-log" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.235358 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-httpd" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.235372 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" containerName="glance-log" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.236051 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.250661 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.343656 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.343689 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.343836 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.343912 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.343990 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344065 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344085 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344113 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344157 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344240 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344299 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbhlk\" (UniqueName: \"kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344349 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344385 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.344418 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445172 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445224 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445247 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445269 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445284 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445304 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445310 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445359 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445324 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445397 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445411 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445436 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445437 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445487 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445564 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445622 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.445665 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbhlk\" (UniqueName: \"kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446097 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446167 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446202 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446275 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446465 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446541 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.446708 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.454255 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.457074 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.474332 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.482416 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbhlk\" (UniqueName: \"kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.493270 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.550664 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.730709 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10fdfbf-dec4-4371-bc6c-a4454f7b37e2" path="/var/lib/kubelet/pods/f10fdfbf-dec4-4371-bc6c-a4454f7b37e2/volumes" Oct 01 09:27:16 crc kubenswrapper[4983]: I1001 09:27:16.953316 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:16 crc kubenswrapper[4983]: W1001 09:27:16.959061 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c4cd69_c879_4f37_b1a0_e92e95170934.slice/crio-cfbfa9b1ff8deae6be4fd5df4cea4b10e2f891bb9fe0cc77fade1477e6554c30 WatchSource:0}: Error finding container cfbfa9b1ff8deae6be4fd5df4cea4b10e2f891bb9fe0cc77fade1477e6554c30: Status 404 returned error can't find the container with id cfbfa9b1ff8deae6be4fd5df4cea4b10e2f891bb9fe0cc77fade1477e6554c30 Oct 01 09:27:17 crc kubenswrapper[4983]: I1001 09:27:17.906226 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerStarted","Data":"a3c4c939499fda01fc6ae7b5fff4232499a1b75ce1c15ebe4c13ef599ed710d7"} Oct 01 09:27:17 crc kubenswrapper[4983]: I1001 09:27:17.908025 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerStarted","Data":"b8d11dd02492eb49c795074314218a77c26dc4e2ea34158afc1cd3ef7ddf0459"} Oct 01 09:27:17 crc kubenswrapper[4983]: I1001 09:27:17.908045 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerStarted","Data":"cfbfa9b1ff8deae6be4fd5df4cea4b10e2f891bb9fe0cc77fade1477e6554c30"} Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.551412 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.551967 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.573153 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.589554 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.601043 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=10.601026062 podStartE2EDuration="10.601026062s" podCreationTimestamp="2025-10-01 09:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:27:17.929993063 +0000 UTC m=+1065.919221860" watchObservedRunningTime="2025-10-01 09:27:26.601026062 +0000 UTC m=+1074.590254859" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.970494 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:26 crc kubenswrapper[4983]: I1001 09:27:26.971616 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:28 crc kubenswrapper[4983]: I1001 09:27:28.884405 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:28 crc kubenswrapper[4983]: I1001 09:27:28.906727 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.497749 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-848kt"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.504832 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-848kt"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.548518 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.548849 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-log" containerID="cri-o://b8d11dd02492eb49c795074314218a77c26dc4e2ea34158afc1cd3ef7ddf0459" gracePeriod=30 Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.548961 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-httpd" containerID="cri-o://a3c4c939499fda01fc6ae7b5fff4232499a1b75ce1c15ebe4c13ef599ed710d7" gracePeriod=30 Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.559749 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.559999 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-log" containerID="cri-o://9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9" gracePeriod=30 Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.560139 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-httpd" containerID="cri-o://eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2" gracePeriod=30 Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.576042 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef9e7-account-delete-kfg5t"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.576892 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.598354 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef9e7-account-delete-kfg5t"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.622827 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-r5bzq"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.629432 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef9e7-account-delete-kfg5t"] Oct 01 09:27:45 crc kubenswrapper[4983]: E1001 09:27:45.630004 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-v4xrl], unattached volumes=[], failed to process volumes=[kube-api-access-v4xrl]: context canceled" pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" podUID="39a6c5a7-6f74-4766-8687-97efbf8d441b" Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.636245 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-r5bzq"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.643656 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f9e7-account-create-gnh4k"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.651136 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f9e7-account-create-gnh4k"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.676155 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4xrl\" (UniqueName: \"kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl\") pod \"glancef9e7-account-delete-kfg5t\" (UID: \"39a6c5a7-6f74-4766-8687-97efbf8d441b\") " pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.700689 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.700940 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="593a0d0b-be96-4c44-8f7e-6741b06554ee" containerName="openstackclient" containerID="cri-o://9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c" gracePeriod=30 Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.777036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4xrl\" (UniqueName: \"kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl\") pod \"glancef9e7-account-delete-kfg5t\" (UID: \"39a6c5a7-6f74-4766-8687-97efbf8d441b\") " pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:45 crc kubenswrapper[4983]: I1001 09:27:45.807380 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4xrl\" (UniqueName: \"kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl\") pod \"glancef9e7-account-delete-kfg5t\" (UID: \"39a6c5a7-6f74-4766-8687-97efbf8d441b\") " pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.038023 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.103161 4983 generic.go:334] "Generic (PLEG): container finished" podID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerID="9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9" exitCode=143 Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.103190 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerDied","Data":"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9"} Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.104411 4983 generic.go:334] "Generic (PLEG): container finished" podID="593a0d0b-be96-4c44-8f7e-6741b06554ee" containerID="9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c" exitCode=143 Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.104450 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.104484 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"593a0d0b-be96-4c44-8f7e-6741b06554ee","Type":"ContainerDied","Data":"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c"} Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.104513 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"593a0d0b-be96-4c44-8f7e-6741b06554ee","Type":"ContainerDied","Data":"5b69bffcb48a0e3f9517170623e1e1f786637979432b0b8b317493ed5fd7d70e"} Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.104528 4983 scope.go:117] "RemoveContainer" containerID="9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.106128 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerID="b8d11dd02492eb49c795074314218a77c26dc4e2ea34158afc1cd3ef7ddf0459" exitCode=143 Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.106179 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.106603 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerDied","Data":"b8d11dd02492eb49c795074314218a77c26dc4e2ea34158afc1cd3ef7ddf0459"} Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.115834 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.122450 4983 scope.go:117] "RemoveContainer" containerID="9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c" Oct 01 09:27:46 crc kubenswrapper[4983]: E1001 09:27:46.122782 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c\": container with ID starting with 9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c not found: ID does not exist" containerID="9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.122838 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c"} err="failed to get container status \"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c\": rpc error: code = NotFound desc = could not find container \"9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c\": container with ID starting with 9e2c5032297742ff5462aa61aa8d065e780a6a954ebaa57f02936240dcf2603c not found: ID does not exist" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.180567 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config\") pod \"593a0d0b-be96-4c44-8f7e-6741b06554ee\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.180668 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ltzj\" (UniqueName: \"kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj\") pod \"593a0d0b-be96-4c44-8f7e-6741b06554ee\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.180751 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts\") pod \"593a0d0b-be96-4c44-8f7e-6741b06554ee\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.180831 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret\") pod \"593a0d0b-be96-4c44-8f7e-6741b06554ee\" (UID: \"593a0d0b-be96-4c44-8f7e-6741b06554ee\") " Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.181954 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "593a0d0b-be96-4c44-8f7e-6741b06554ee" (UID: "593a0d0b-be96-4c44-8f7e-6741b06554ee"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.185557 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj" (OuterVolumeSpecName: "kube-api-access-7ltzj") pod "593a0d0b-be96-4c44-8f7e-6741b06554ee" (UID: "593a0d0b-be96-4c44-8f7e-6741b06554ee"). InnerVolumeSpecName "kube-api-access-7ltzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.199617 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "593a0d0b-be96-4c44-8f7e-6741b06554ee" (UID: "593a0d0b-be96-4c44-8f7e-6741b06554ee"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.200511 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "593a0d0b-be96-4c44-8f7e-6741b06554ee" (UID: "593a0d0b-be96-4c44-8f7e-6741b06554ee"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.282402 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4xrl\" (UniqueName: \"kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl\") pod \"39a6c5a7-6f74-4766-8687-97efbf8d441b\" (UID: \"39a6c5a7-6f74-4766-8687-97efbf8d441b\") " Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.282780 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ltzj\" (UniqueName: \"kubernetes.io/projected/593a0d0b-be96-4c44-8f7e-6741b06554ee-kube-api-access-7ltzj\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.282794 4983 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.282802 4983 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.282843 4983 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/593a0d0b-be96-4c44-8f7e-6741b06554ee-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.284906 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl" (OuterVolumeSpecName: "kube-api-access-v4xrl") pod "39a6c5a7-6f74-4766-8687-97efbf8d441b" (UID: "39a6c5a7-6f74-4766-8687-97efbf8d441b"). InnerVolumeSpecName "kube-api-access-v4xrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.383800 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4xrl\" (UniqueName: \"kubernetes.io/projected/39a6c5a7-6f74-4766-8687-97efbf8d441b-kube-api-access-v4xrl\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.432934 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.438397 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.722834 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4de3d787-7a87-4578-ad12-3e338a474309" path="/var/lib/kubelet/pods/4de3d787-7a87-4578-ad12-3e338a474309/volumes" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.723597 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593a0d0b-be96-4c44-8f7e-6741b06554ee" path="/var/lib/kubelet/pods/593a0d0b-be96-4c44-8f7e-6741b06554ee/volumes" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.724046 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4c026f-de4c-4a34-af25-a41370a5af55" path="/var/lib/kubelet/pods/9d4c026f-de4c-4a34-af25-a41370a5af55/volumes" Oct 01 09:27:46 crc kubenswrapper[4983]: I1001 09:27:46.724551 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4091bd7-099e-4663-abdd-189b8131920b" path="/var/lib/kubelet/pods/a4091bd7-099e-4663-abdd-189b8131920b/volumes" Oct 01 09:27:47 crc kubenswrapper[4983]: I1001 09:27:47.112450 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef9e7-account-delete-kfg5t" Oct 01 09:27:47 crc kubenswrapper[4983]: I1001 09:27:47.145985 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef9e7-account-delete-kfg5t"] Oct 01 09:27:47 crc kubenswrapper[4983]: I1001 09:27:47.149839 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef9e7-account-delete-kfg5t"] Oct 01 09:27:48 crc kubenswrapper[4983]: I1001 09:27:48.722686 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39a6c5a7-6f74-4766-8687-97efbf8d441b" path="/var/lib/kubelet/pods/39a6c5a7-6f74-4766-8687-97efbf8d441b/volumes" Oct 01 09:27:48 crc kubenswrapper[4983]: I1001 09:27:48.958152 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": read tcp 10.217.0.2:55212->10.217.0.103:9292: read: connection reset by peer" Oct 01 09:27:48 crc kubenswrapper[4983]: I1001 09:27:48.958185 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": read tcp 10.217.0.2:55220->10.217.0.103:9292: read: connection reset by peer" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.102300 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.129114 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerID="a3c4c939499fda01fc6ae7b5fff4232499a1b75ce1c15ebe4c13ef599ed710d7" exitCode=0 Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.129219 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerDied","Data":"a3c4c939499fda01fc6ae7b5fff4232499a1b75ce1c15ebe4c13ef599ed710d7"} Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.131525 4983 generic.go:334] "Generic (PLEG): container finished" podID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerID="eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2" exitCode=0 Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.131554 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerDied","Data":"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2"} Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.131596 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce","Type":"ContainerDied","Data":"5357d795c28ed14c98d82cb4da5a20f4f31b9b9df8353a1c6cc731629ad6288f"} Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.131604 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.131624 4983 scope.go:117] "RemoveContainer" containerID="eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.151270 4983 scope.go:117] "RemoveContainer" containerID="9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.176781 4983 scope.go:117] "RemoveContainer" containerID="eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2" Oct 01 09:27:49 crc kubenswrapper[4983]: E1001 09:27:49.177332 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2\": container with ID starting with eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2 not found: ID does not exist" containerID="eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.177382 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2"} err="failed to get container status \"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2\": rpc error: code = NotFound desc = could not find container \"eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2\": container with ID starting with eda813ad12e2d201d384f4e6ce75d41b0384e8664394f8f0f4d3ad68f3dcc6b2 not found: ID does not exist" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.177408 4983 scope.go:117] "RemoveContainer" containerID="9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9" Oct 01 09:27:49 crc kubenswrapper[4983]: E1001 09:27:49.177934 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9\": container with ID starting with 9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9 not found: ID does not exist" containerID="9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.177977 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9"} err="failed to get container status \"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9\": rpc error: code = NotFound desc = could not find container \"9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9\": container with ID starting with 9481e2bf0dc0e691d4b110eb65ec82ab07da23986b288c9389c9fed53e2b7dc9 not found: ID does not exist" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228717 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228761 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228795 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228832 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228877 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228909 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228929 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228921 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.228991 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run" (OuterVolumeSpecName: "run") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229016 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229036 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229063 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229067 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xft56\" (UniqueName: \"kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229109 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229145 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229171 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229200 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229234 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run\") pod \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\" (UID: \"c79d04e6-6bf4-4baa-a6b3-cce69ae405ce\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229562 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs" (OuterVolumeSpecName: "logs") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229615 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev" (OuterVolumeSpecName: "dev") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229652 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys" (OuterVolumeSpecName: "sys") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229679 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.229997 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230017 4983 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-dev\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230043 4983 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-sys\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230064 4983 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230083 4983 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230101 4983 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230116 4983 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230132 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.230148 4983 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.237397 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56" (OuterVolumeSpecName: "kube-api-access-xft56") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "kube-api-access-xft56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.237589 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.243755 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts" (OuterVolumeSpecName: "scripts") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.245185 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.284654 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data" (OuterVolumeSpecName: "config-data") pod "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" (UID: "c79d04e6-6bf4-4baa-a6b3-cce69ae405ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.327068 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331184 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331214 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331245 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331283 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331307 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xft56\" (UniqueName: \"kubernetes.io/projected/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce-kube-api-access-xft56\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.331322 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.344048 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.344075 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432289 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432372 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432421 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432465 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432480 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432500 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbhlk\" (UniqueName: \"kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432561 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432579 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432594 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432660 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432689 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432709 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432753 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs\") pod \"d4c4cd69-c879-4f37-b1a0-e92e95170934\" (UID: \"d4c4cd69-c879-4f37-b1a0-e92e95170934\") " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432799 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys" (OuterVolumeSpecName: "sys") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432862 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432863 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run" (OuterVolumeSpecName: "run") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432839 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432826 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432945 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.432983 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev" (OuterVolumeSpecName: "dev") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433363 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433532 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs" (OuterVolumeSpecName: "logs") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433535 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433573 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433587 4983 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433600 4983 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433611 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433623 4983 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-sys\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433633 4983 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433643 4983 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433657 4983 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-dev\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.433669 4983 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4c4cd69-c879-4f37-b1a0-e92e95170934-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.435501 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.435622 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts" (OuterVolumeSpecName: "scripts") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.435964 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.436300 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk" (OuterVolumeSpecName: "kube-api-access-sbhlk") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "kube-api-access-sbhlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.461168 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.465767 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.468415 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data" (OuterVolumeSpecName: "config-data") pod "d4c4cd69-c879-4f37-b1a0-e92e95170934" (UID: "d4c4cd69-c879-4f37-b1a0-e92e95170934"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535423 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535466 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbhlk\" (UniqueName: \"kubernetes.io/projected/d4c4cd69-c879-4f37-b1a0-e92e95170934-kube-api-access-sbhlk\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535536 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535549 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4c4cd69-c879-4f37-b1a0-e92e95170934-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535567 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.535581 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c4cd69-c879-4f37-b1a0-e92e95170934-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.547159 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.548508 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.636829 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:49 crc kubenswrapper[4983]: I1001 09:27:49.636858 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.139045 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d4c4cd69-c879-4f37-b1a0-e92e95170934","Type":"ContainerDied","Data":"cfbfa9b1ff8deae6be4fd5df4cea4b10e2f891bb9fe0cc77fade1477e6554c30"} Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.139105 4983 scope.go:117] "RemoveContainer" containerID="a3c4c939499fda01fc6ae7b5fff4232499a1b75ce1c15ebe4c13ef599ed710d7" Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.139140 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.160908 4983 scope.go:117] "RemoveContainer" containerID="b8d11dd02492eb49c795074314218a77c26dc4e2ea34158afc1cd3ef7ddf0459" Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.165090 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.169604 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.723422 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" path="/var/lib/kubelet/pods/c79d04e6-6bf4-4baa-a6b3-cce69ae405ce/volumes" Oct 01 09:27:50 crc kubenswrapper[4983]: I1001 09:27:50.724346 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" path="/var/lib/kubelet/pods/d4c4cd69-c879-4f37-b1a0-e92e95170934/volumes" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100019 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-c2nm8"] Oct 01 09:27:51 crc kubenswrapper[4983]: E1001 09:27:51.100286 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100297 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: E1001 09:27:51.100348 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100355 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: E1001 09:27:51.100370 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100376 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: E1001 09:27:51.100387 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593a0d0b-be96-4c44-8f7e-6741b06554ee" containerName="openstackclient" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100392 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="593a0d0b-be96-4c44-8f7e-6741b06554ee" containerName="openstackclient" Oct 01 09:27:51 crc kubenswrapper[4983]: E1001 09:27:51.100400 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100406 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100538 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="593a0d0b-be96-4c44-8f7e-6741b06554ee" containerName="openstackclient" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100551 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100558 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79d04e6-6bf4-4baa-a6b3-cce69ae405ce" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100570 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-log" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.100582 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c4cd69-c879-4f37-b1a0-e92e95170934" containerName="glance-httpd" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.101099 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.114908 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-c2nm8"] Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.256482 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfs9q\" (UniqueName: \"kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q\") pod \"glance-db-create-c2nm8\" (UID: \"aa07727a-309c-41a5-8bd5-d8346c072a60\") " pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.357768 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfs9q\" (UniqueName: \"kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q\") pod \"glance-db-create-c2nm8\" (UID: \"aa07727a-309c-41a5-8bd5-d8346c072a60\") " pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.374261 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfs9q\" (UniqueName: \"kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q\") pod \"glance-db-create-c2nm8\" (UID: \"aa07727a-309c-41a5-8bd5-d8346c072a60\") " pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.416155 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:51 crc kubenswrapper[4983]: I1001 09:27:51.904045 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-c2nm8"] Oct 01 09:27:52 crc kubenswrapper[4983]: I1001 09:27:52.156473 4983 generic.go:334] "Generic (PLEG): container finished" podID="aa07727a-309c-41a5-8bd5-d8346c072a60" containerID="c20f42e2eae620159a19d0e70c13310bd9f15f46a9e185bb78f4e39ebbfa72b2" exitCode=0 Oct 01 09:27:52 crc kubenswrapper[4983]: I1001 09:27:52.156517 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-c2nm8" event={"ID":"aa07727a-309c-41a5-8bd5-d8346c072a60","Type":"ContainerDied","Data":"c20f42e2eae620159a19d0e70c13310bd9f15f46a9e185bb78f4e39ebbfa72b2"} Oct 01 09:27:52 crc kubenswrapper[4983]: I1001 09:27:52.156546 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-c2nm8" event={"ID":"aa07727a-309c-41a5-8bd5-d8346c072a60","Type":"ContainerStarted","Data":"2b0974677b85a8cf67f5d8ba142b5c32e09ca0f505d9b3dc060ee1ce0a7ed773"} Oct 01 09:27:53 crc kubenswrapper[4983]: I1001 09:27:53.446531 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:27:53 crc kubenswrapper[4983]: I1001 09:27:53.590297 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfs9q\" (UniqueName: \"kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q\") pod \"aa07727a-309c-41a5-8bd5-d8346c072a60\" (UID: \"aa07727a-309c-41a5-8bd5-d8346c072a60\") " Oct 01 09:27:53 crc kubenswrapper[4983]: I1001 09:27:53.595168 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q" (OuterVolumeSpecName: "kube-api-access-pfs9q") pod "aa07727a-309c-41a5-8bd5-d8346c072a60" (UID: "aa07727a-309c-41a5-8bd5-d8346c072a60"). InnerVolumeSpecName "kube-api-access-pfs9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:27:53 crc kubenswrapper[4983]: I1001 09:27:53.691608 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfs9q\" (UniqueName: \"kubernetes.io/projected/aa07727a-309c-41a5-8bd5-d8346c072a60-kube-api-access-pfs9q\") on node \"crc\" DevicePath \"\"" Oct 01 09:27:54 crc kubenswrapper[4983]: I1001 09:27:54.169759 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-c2nm8" event={"ID":"aa07727a-309c-41a5-8bd5-d8346c072a60","Type":"ContainerDied","Data":"2b0974677b85a8cf67f5d8ba142b5c32e09ca0f505d9b3dc060ee1ce0a7ed773"} Oct 01 09:27:54 crc kubenswrapper[4983]: I1001 09:27:54.169830 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b0974677b85a8cf67f5d8ba142b5c32e09ca0f505d9b3dc060ee1ce0a7ed773" Oct 01 09:27:54 crc kubenswrapper[4983]: I1001 09:27:54.169837 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-c2nm8" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.222094 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-196f-account-create-pdlcd"] Oct 01 09:28:01 crc kubenswrapper[4983]: E1001 09:28:01.222895 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa07727a-309c-41a5-8bd5-d8346c072a60" containerName="mariadb-database-create" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.222910 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa07727a-309c-41a5-8bd5-d8346c072a60" containerName="mariadb-database-create" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.223087 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa07727a-309c-41a5-8bd5-d8346c072a60" containerName="mariadb-database-create" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.223582 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.226370 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.234478 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-196f-account-create-pdlcd"] Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.396803 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w22zh\" (UniqueName: \"kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh\") pod \"glance-196f-account-create-pdlcd\" (UID: \"ebfaf138-4fa4-4a92-b766-91f7f3d57825\") " pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.497715 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w22zh\" (UniqueName: \"kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh\") pod \"glance-196f-account-create-pdlcd\" (UID: \"ebfaf138-4fa4-4a92-b766-91f7f3d57825\") " pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.528701 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w22zh\" (UniqueName: \"kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh\") pod \"glance-196f-account-create-pdlcd\" (UID: \"ebfaf138-4fa4-4a92-b766-91f7f3d57825\") " pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.541190 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:01 crc kubenswrapper[4983]: I1001 09:28:01.754422 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-196f-account-create-pdlcd"] Oct 01 09:28:02 crc kubenswrapper[4983]: I1001 09:28:02.231995 4983 generic.go:334] "Generic (PLEG): container finished" podID="ebfaf138-4fa4-4a92-b766-91f7f3d57825" containerID="7cfd1c70f4e00a448514c47f43c1ea095b6f4f35d522652294041ed21f74214a" exitCode=0 Oct 01 09:28:02 crc kubenswrapper[4983]: I1001 09:28:02.232077 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" event={"ID":"ebfaf138-4fa4-4a92-b766-91f7f3d57825","Type":"ContainerDied","Data":"7cfd1c70f4e00a448514c47f43c1ea095b6f4f35d522652294041ed21f74214a"} Oct 01 09:28:02 crc kubenswrapper[4983]: I1001 09:28:02.232372 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" event={"ID":"ebfaf138-4fa4-4a92-b766-91f7f3d57825","Type":"ContainerStarted","Data":"a4831c53fe139c1ea52973232659f0b39c4f0838078e3dcd95d3789121707d30"} Oct 01 09:28:03 crc kubenswrapper[4983]: I1001 09:28:03.486914 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:03 crc kubenswrapper[4983]: I1001 09:28:03.626177 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w22zh\" (UniqueName: \"kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh\") pod \"ebfaf138-4fa4-4a92-b766-91f7f3d57825\" (UID: \"ebfaf138-4fa4-4a92-b766-91f7f3d57825\") " Oct 01 09:28:03 crc kubenswrapper[4983]: I1001 09:28:03.631854 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh" (OuterVolumeSpecName: "kube-api-access-w22zh") pod "ebfaf138-4fa4-4a92-b766-91f7f3d57825" (UID: "ebfaf138-4fa4-4a92-b766-91f7f3d57825"). InnerVolumeSpecName "kube-api-access-w22zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:03 crc kubenswrapper[4983]: I1001 09:28:03.727709 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w22zh\" (UniqueName: \"kubernetes.io/projected/ebfaf138-4fa4-4a92-b766-91f7f3d57825-kube-api-access-w22zh\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:04 crc kubenswrapper[4983]: I1001 09:28:04.254626 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" event={"ID":"ebfaf138-4fa4-4a92-b766-91f7f3d57825","Type":"ContainerDied","Data":"a4831c53fe139c1ea52973232659f0b39c4f0838078e3dcd95d3789121707d30"} Oct 01 09:28:04 crc kubenswrapper[4983]: I1001 09:28:04.254673 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4831c53fe139c1ea52973232659f0b39c4f0838078e3dcd95d3789121707d30" Oct 01 09:28:04 crc kubenswrapper[4983]: I1001 09:28:04.254732 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-196f-account-create-pdlcd" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.365910 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-flwfn"] Oct 01 09:28:06 crc kubenswrapper[4983]: E1001 09:28:06.366496 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebfaf138-4fa4-4a92-b766-91f7f3d57825" containerName="mariadb-account-create" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.366512 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebfaf138-4fa4-4a92-b766-91f7f3d57825" containerName="mariadb-account-create" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.366716 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebfaf138-4fa4-4a92-b766-91f7f3d57825" containerName="mariadb-account-create" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.367248 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.370621 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.370680 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-m9w54" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.371273 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.388587 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-flwfn"] Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.569444 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.569500 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9hng\" (UniqueName: \"kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.569543 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.569580 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.671726 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.671903 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9hng\" (UniqueName: \"kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.671996 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.672075 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.678184 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.680307 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.681155 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.692139 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9hng\" (UniqueName: \"kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng\") pod \"glance-db-sync-flwfn\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:06 crc kubenswrapper[4983]: I1001 09:28:06.984987 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:07 crc kubenswrapper[4983]: I1001 09:28:07.371737 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-flwfn"] Oct 01 09:28:08 crc kubenswrapper[4983]: I1001 09:28:08.282560 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-flwfn" event={"ID":"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00","Type":"ContainerStarted","Data":"ca96ad9198679e8229510fd6a04d3c26231d6483d63facb7d237d74927488021"} Oct 01 09:28:08 crc kubenswrapper[4983]: I1001 09:28:08.283006 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-flwfn" event={"ID":"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00","Type":"ContainerStarted","Data":"a7ae9dfbfa308e2c01ea4d720e1a9a791cafb1e3dfd7028780b03519993db7f0"} Oct 01 09:28:11 crc kubenswrapper[4983]: I1001 09:28:11.306111 4983 generic.go:334] "Generic (PLEG): container finished" podID="4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" containerID="ca96ad9198679e8229510fd6a04d3c26231d6483d63facb7d237d74927488021" exitCode=0 Oct 01 09:28:11 crc kubenswrapper[4983]: I1001 09:28:11.306224 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-flwfn" event={"ID":"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00","Type":"ContainerDied","Data":"ca96ad9198679e8229510fd6a04d3c26231d6483d63facb7d237d74927488021"} Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.674879 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.859839 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle\") pod \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.859895 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9hng\" (UniqueName: \"kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng\") pod \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.859946 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data\") pod \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.859981 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data\") pod \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\" (UID: \"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00\") " Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.865050 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng" (OuterVolumeSpecName: "kube-api-access-f9hng") pod "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" (UID: "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00"). InnerVolumeSpecName "kube-api-access-f9hng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.865344 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" (UID: "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.882269 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" (UID: "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.892980 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data" (OuterVolumeSpecName: "config-data") pod "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" (UID: "4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.962659 4983 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.962753 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9hng\" (UniqueName: \"kubernetes.io/projected/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-kube-api-access-f9hng\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.962784 4983 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:12 crc kubenswrapper[4983]: I1001 09:28:12.962858 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.334138 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-flwfn" event={"ID":"4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00","Type":"ContainerDied","Data":"a7ae9dfbfa308e2c01ea4d720e1a9a791cafb1e3dfd7028780b03519993db7f0"} Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.334219 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-flwfn" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.334233 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ae9dfbfa308e2c01ea4d720e1a9a791cafb1e3dfd7028780b03519993db7f0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.584778 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:13 crc kubenswrapper[4983]: E1001 09:28:13.589188 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" containerName="glance-db-sync" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.589225 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" containerName="glance-db-sync" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.589424 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" containerName="glance-db-sync" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.590298 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.595577 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.596392 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.596573 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.596475 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.597097 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-m9w54" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.598574 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.601353 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.611936 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:13 crc kubenswrapper[4983]: E1001 09:28:13.612439 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-7ntm6 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-7ntm6 logs public-tls-certs scripts]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="402615f4-ce9f-4b17-a9dd-93c8376d3af1" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.771238 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.771865 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.771985 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772120 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ntm6\" (UniqueName: \"kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772155 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772231 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772261 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772284 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.772343 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874109 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874155 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874177 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874213 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874261 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874280 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874307 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874340 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ntm6\" (UniqueName: \"kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.874356 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.875175 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.875210 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.875481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.878522 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.878879 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.879227 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.880162 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.890968 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.893387 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ntm6\" (UniqueName: \"kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:13 crc kubenswrapper[4983]: I1001 09:28:13.896425 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.341149 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.350553 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483415 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483495 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483572 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483603 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483634 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483680 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483705 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483726 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ntm6\" (UniqueName: \"kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483746 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs\") pod \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\" (UID: \"402615f4-ce9f-4b17-a9dd-93c8376d3af1\") " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.483746 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs" (OuterVolumeSpecName: "logs") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.484075 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.484081 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.487839 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.488124 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts" (OuterVolumeSpecName: "scripts") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.488352 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.488418 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data" (OuterVolumeSpecName: "config-data") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.488653 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6" (OuterVolumeSpecName: "kube-api-access-7ntm6") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "kube-api-access-7ntm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.488783 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.489069 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "402615f4-ce9f-4b17-a9dd-93c8376d3af1" (UID: "402615f4-ce9f-4b17-a9dd-93c8376d3af1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.585934 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/402615f4-ce9f-4b17-a9dd-93c8376d3af1-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.585977 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ntm6\" (UniqueName: \"kubernetes.io/projected/402615f4-ce9f-4b17-a9dd-93c8376d3af1-kube-api-access-7ntm6\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.585988 4983 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.585997 4983 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.586006 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.586015 4983 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.586045 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.586054 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/402615f4-ce9f-4b17-a9dd-93c8376d3af1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.598203 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:28:14 crc kubenswrapper[4983]: I1001 09:28:14.687650 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:15 crc kubenswrapper[4983]: I1001 09:28:15.349432 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:15 crc kubenswrapper[4983]: I1001 09:28:15.382358 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:15 crc kubenswrapper[4983]: I1001 09:28:15.398542 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:16 crc kubenswrapper[4983]: I1001 09:28:16.727553 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="402615f4-ce9f-4b17-a9dd-93c8376d3af1" path="/var/lib/kubelet/pods/402615f4-ce9f-4b17-a9dd-93c8376d3af1/volumes" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.197698 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.199724 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.202546 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-m9w54" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.202651 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.202667 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.202825 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.205373 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.205537 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.210098 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.363906 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.363991 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364019 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364060 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364085 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364105 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364125 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364140 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.364316 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcvwd\" (UniqueName: \"kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466179 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466751 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466851 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466890 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466931 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.466972 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467014 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467045 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467117 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcvwd\" (UniqueName: \"kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467169 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467472 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.467619 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.474298 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.474413 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.475318 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.475825 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.476439 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.489985 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcvwd\" (UniqueName: \"kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.491877 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.519222 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:20 crc kubenswrapper[4983]: I1001 09:28:20.956574 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:21 crc kubenswrapper[4983]: I1001 09:28:21.395698 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerStarted","Data":"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8"} Oct 01 09:28:21 crc kubenswrapper[4983]: I1001 09:28:21.395748 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerStarted","Data":"9fe86308230311592236352e8212c71d44107873f81d9d84e82f5d0f6174ad64"} Oct 01 09:28:22 crc kubenswrapper[4983]: I1001 09:28:22.405453 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerStarted","Data":"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787"} Oct 01 09:28:22 crc kubenswrapper[4983]: I1001 09:28:22.432857 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.432843836 podStartE2EDuration="2.432843836s" podCreationTimestamp="2025-10-01 09:28:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:28:22.43004627 +0000 UTC m=+1130.419275077" watchObservedRunningTime="2025-10-01 09:28:22.432843836 +0000 UTC m=+1130.422072633" Oct 01 09:28:30 crc kubenswrapper[4983]: I1001 09:28:30.519730 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:30 crc kubenswrapper[4983]: I1001 09:28:30.520295 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:30 crc kubenswrapper[4983]: I1001 09:28:30.541968 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:30 crc kubenswrapper[4983]: I1001 09:28:30.566945 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:31 crc kubenswrapper[4983]: I1001 09:28:31.478309 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:31 crc kubenswrapper[4983]: I1001 09:28:31.478361 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:33 crc kubenswrapper[4983]: I1001 09:28:33.402759 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:33 crc kubenswrapper[4983]: I1001 09:28:33.404865 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.107577 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-flwfn"] Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.115239 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-flwfn"] Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.180053 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance196f-account-delete-rq46k"] Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.195229 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.195378 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.205460 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance196f-account-delete-rq46k"] Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.280703 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxvt7\" (UniqueName: \"kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7\") pod \"glance196f-account-delete-rq46k\" (UID: \"e18885bd-316e-4101-a443-f2b82a0776d0\") " pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.381586 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxvt7\" (UniqueName: \"kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7\") pod \"glance196f-account-delete-rq46k\" (UID: \"e18885bd-316e-4101-a443-f2b82a0776d0\") " pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.398575 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxvt7\" (UniqueName: \"kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7\") pod \"glance196f-account-delete-rq46k\" (UID: \"e18885bd-316e-4101-a443-f2b82a0776d0\") " pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.502210 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-httpd" containerID="cri-o://7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787" gracePeriod=30 Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.502133 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-log" containerID="cri-o://9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8" gracePeriod=30 Oct 01 09:28:35 crc kubenswrapper[4983]: I1001 09:28:35.530441 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:35.955181 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance196f-account-delete-rq46k"] Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.509485 4983 generic.go:334] "Generic (PLEG): container finished" podID="e18885bd-316e-4101-a443-f2b82a0776d0" containerID="6040c8cf1f30a295769765f93d561aa693a90d3fe36f695993c499bcc27d75bc" exitCode=0 Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.509585 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" event={"ID":"e18885bd-316e-4101-a443-f2b82a0776d0","Type":"ContainerDied","Data":"6040c8cf1f30a295769765f93d561aa693a90d3fe36f695993c499bcc27d75bc"} Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.509621 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" event={"ID":"e18885bd-316e-4101-a443-f2b82a0776d0","Type":"ContainerStarted","Data":"75f26dcfc76c1e3e7026392409567905bb5df177334cddfb2601c39181794390"} Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.512189 4983 generic.go:334] "Generic (PLEG): container finished" podID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerID="9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8" exitCode=143 Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.512234 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerDied","Data":"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8"} Oct 01 09:28:36 crc kubenswrapper[4983]: I1001 09:28:36.725369 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00" path="/var/lib/kubelet/pods/4bbd8ca2-c51f-4cbe-bafd-c7e8ffd2fd00/volumes" Oct 01 09:28:37 crc kubenswrapper[4983]: I1001 09:28:37.756308 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:37 crc kubenswrapper[4983]: I1001 09:28:37.917277 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxvt7\" (UniqueName: \"kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7\") pod \"e18885bd-316e-4101-a443-f2b82a0776d0\" (UID: \"e18885bd-316e-4101-a443-f2b82a0776d0\") " Oct 01 09:28:37 crc kubenswrapper[4983]: I1001 09:28:37.921835 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7" (OuterVolumeSpecName: "kube-api-access-mxvt7") pod "e18885bd-316e-4101-a443-f2b82a0776d0" (UID: "e18885bd-316e-4101-a443-f2b82a0776d0"). InnerVolumeSpecName "kube-api-access-mxvt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:38 crc kubenswrapper[4983]: I1001 09:28:38.019566 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxvt7\" (UniqueName: \"kubernetes.io/projected/e18885bd-316e-4101-a443-f2b82a0776d0-kube-api-access-mxvt7\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:38 crc kubenswrapper[4983]: I1001 09:28:38.526541 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" event={"ID":"e18885bd-316e-4101-a443-f2b82a0776d0","Type":"ContainerDied","Data":"75f26dcfc76c1e3e7026392409567905bb5df177334cddfb2601c39181794390"} Oct 01 09:28:38 crc kubenswrapper[4983]: I1001 09:28:38.526573 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance196f-account-delete-rq46k" Oct 01 09:28:38 crc kubenswrapper[4983]: I1001 09:28:38.526581 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f26dcfc76c1e3e7026392409567905bb5df177334cddfb2601c39181794390" Oct 01 09:28:38 crc kubenswrapper[4983]: I1001 09:28:38.885408 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.032990 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033036 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033092 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcvwd\" (UniqueName: \"kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033159 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033177 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033247 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033834 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033889 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7579683b-d77c-40e8-8254-b88fb2a2818b\" (UID: \"7579683b-d77c-40e8-8254-b88fb2a2818b\") " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.033877 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.034231 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.034243 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs" (OuterVolumeSpecName: "logs") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.037278 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.037775 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd" (OuterVolumeSpecName: "kube-api-access-tcvwd") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "kube-api-access-tcvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.037936 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts" (OuterVolumeSpecName: "scripts") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.051963 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.065330 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.069245 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.072498 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data" (OuterVolumeSpecName: "config-data") pod "7579683b-d77c-40e8-8254-b88fb2a2818b" (UID: "7579683b-d77c-40e8-8254-b88fb2a2818b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135628 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7579683b-d77c-40e8-8254-b88fb2a2818b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135667 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135680 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcvwd\" (UniqueName: \"kubernetes.io/projected/7579683b-d77c-40e8-8254-b88fb2a2818b-kube-api-access-tcvwd\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135694 4983 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135706 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135717 4983 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135727 4983 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7579683b-d77c-40e8-8254-b88fb2a2818b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.135760 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.149340 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.236926 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.535276 4983 generic.go:334] "Generic (PLEG): container finished" podID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerID="7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787" exitCode=0 Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.535318 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerDied","Data":"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787"} Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.535364 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7579683b-d77c-40e8-8254-b88fb2a2818b","Type":"ContainerDied","Data":"9fe86308230311592236352e8212c71d44107873f81d9d84e82f5d0f6174ad64"} Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.535383 4983 scope.go:117] "RemoveContainer" containerID="7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.535396 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.562584 4983 scope.go:117] "RemoveContainer" containerID="9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.570303 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.577551 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.585331 4983 scope.go:117] "RemoveContainer" containerID="7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787" Oct 01 09:28:39 crc kubenswrapper[4983]: E1001 09:28:39.585781 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787\": container with ID starting with 7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787 not found: ID does not exist" containerID="7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.585836 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787"} err="failed to get container status \"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787\": rpc error: code = NotFound desc = could not find container \"7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787\": container with ID starting with 7f25d1bf21e4de5dc3b287bae35fad4a5728e93ce1be8345474041ff459a0787 not found: ID does not exist" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.585865 4983 scope.go:117] "RemoveContainer" containerID="9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8" Oct 01 09:28:39 crc kubenswrapper[4983]: E1001 09:28:39.586218 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8\": container with ID starting with 9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8 not found: ID does not exist" containerID="9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8" Oct 01 09:28:39 crc kubenswrapper[4983]: I1001 09:28:39.586241 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8"} err="failed to get container status \"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8\": rpc error: code = NotFound desc = could not find container \"9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8\": container with ID starting with 9b8804aafb92a2a46dd9fd5fc96af5d8654aa32c7b9a95b23c960a9b06f6d5b8 not found: ID does not exist" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.184794 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-c2nm8"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.199404 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-c2nm8"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.207534 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-196f-account-create-pdlcd"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.213643 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance196f-account-delete-rq46k"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.221097 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance196f-account-delete-rq46k"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.225551 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-196f-account-create-pdlcd"] Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.726923 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" path="/var/lib/kubelet/pods/7579683b-d77c-40e8-8254-b88fb2a2818b/volumes" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.728196 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa07727a-309c-41a5-8bd5-d8346c072a60" path="/var/lib/kubelet/pods/aa07727a-309c-41a5-8bd5-d8346c072a60/volumes" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.729401 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e18885bd-316e-4101-a443-f2b82a0776d0" path="/var/lib/kubelet/pods/e18885bd-316e-4101-a443-f2b82a0776d0/volumes" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.730996 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebfaf138-4fa4-4a92-b766-91f7f3d57825" path="/var/lib/kubelet/pods/ebfaf138-4fa4-4a92-b766-91f7f3d57825/volumes" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894187 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-zgwdz"] Oct 01 09:28:40 crc kubenswrapper[4983]: E1001 09:28:40.894569 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18885bd-316e-4101-a443-f2b82a0776d0" containerName="mariadb-account-delete" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894592 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18885bd-316e-4101-a443-f2b82a0776d0" containerName="mariadb-account-delete" Oct 01 09:28:40 crc kubenswrapper[4983]: E1001 09:28:40.894615 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-httpd" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894624 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-httpd" Oct 01 09:28:40 crc kubenswrapper[4983]: E1001 09:28:40.894638 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-log" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894679 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-log" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894861 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-log" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894885 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18885bd-316e-4101-a443-f2b82a0776d0" containerName="mariadb-account-delete" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.894897 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="7579683b-d77c-40e8-8254-b88fb2a2818b" containerName="glance-httpd" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.895542 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:40 crc kubenswrapper[4983]: I1001 09:28:40.902358 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zgwdz"] Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.065111 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltw99\" (UniqueName: \"kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99\") pod \"glance-db-create-zgwdz\" (UID: \"f0301a58-c508-4ba4-8c7a-d0e3f845bba9\") " pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.166607 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltw99\" (UniqueName: \"kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99\") pod \"glance-db-create-zgwdz\" (UID: \"f0301a58-c508-4ba4-8c7a-d0e3f845bba9\") " pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.187577 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltw99\" (UniqueName: \"kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99\") pod \"glance-db-create-zgwdz\" (UID: \"f0301a58-c508-4ba4-8c7a-d0e3f845bba9\") " pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.225336 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.491546 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zgwdz"] Oct 01 09:28:41 crc kubenswrapper[4983]: W1001 09:28:41.496746 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0301a58_c508_4ba4_8c7a_d0e3f845bba9.slice/crio-f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f WatchSource:0}: Error finding container f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f: Status 404 returned error can't find the container with id f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f Oct 01 09:28:41 crc kubenswrapper[4983]: I1001 09:28:41.553939 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zgwdz" event={"ID":"f0301a58-c508-4ba4-8c7a-d0e3f845bba9","Type":"ContainerStarted","Data":"f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f"} Oct 01 09:28:42 crc kubenswrapper[4983]: I1001 09:28:42.564213 4983 generic.go:334] "Generic (PLEG): container finished" podID="f0301a58-c508-4ba4-8c7a-d0e3f845bba9" containerID="5f7cf39ad20a89f71338bf4c2cca4900fd24a57352121b6ab5f44e8af7049ecf" exitCode=0 Oct 01 09:28:42 crc kubenswrapper[4983]: I1001 09:28:42.564270 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zgwdz" event={"ID":"f0301a58-c508-4ba4-8c7a-d0e3f845bba9","Type":"ContainerDied","Data":"5f7cf39ad20a89f71338bf4c2cca4900fd24a57352121b6ab5f44e8af7049ecf"} Oct 01 09:28:43 crc kubenswrapper[4983]: I1001 09:28:43.828066 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:43 crc kubenswrapper[4983]: I1001 09:28:43.906150 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltw99\" (UniqueName: \"kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99\") pod \"f0301a58-c508-4ba4-8c7a-d0e3f845bba9\" (UID: \"f0301a58-c508-4ba4-8c7a-d0e3f845bba9\") " Oct 01 09:28:43 crc kubenswrapper[4983]: I1001 09:28:43.915054 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99" (OuterVolumeSpecName: "kube-api-access-ltw99") pod "f0301a58-c508-4ba4-8c7a-d0e3f845bba9" (UID: "f0301a58-c508-4ba4-8c7a-d0e3f845bba9"). InnerVolumeSpecName "kube-api-access-ltw99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:44 crc kubenswrapper[4983]: I1001 09:28:44.007934 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltw99\" (UniqueName: \"kubernetes.io/projected/f0301a58-c508-4ba4-8c7a-d0e3f845bba9-kube-api-access-ltw99\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:44 crc kubenswrapper[4983]: I1001 09:28:44.582053 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zgwdz" event={"ID":"f0301a58-c508-4ba4-8c7a-d0e3f845bba9","Type":"ContainerDied","Data":"f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f"} Oct 01 09:28:44 crc kubenswrapper[4983]: I1001 09:28:44.582096 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f26d5e8ee8cc0b181e0c6889444dcffea8e3bc9628b5631338f2138ad1bdb19f" Oct 01 09:28:44 crc kubenswrapper[4983]: I1001 09:28:44.582137 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zgwdz" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.922680 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-c322-account-create-ttpj4"] Oct 01 09:28:50 crc kubenswrapper[4983]: E1001 09:28:50.923542 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0301a58-c508-4ba4-8c7a-d0e3f845bba9" containerName="mariadb-database-create" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.923579 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0301a58-c508-4ba4-8c7a-d0e3f845bba9" containerName="mariadb-database-create" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.923776 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0301a58-c508-4ba4-8c7a-d0e3f845bba9" containerName="mariadb-database-create" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.924344 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.927841 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 01 09:28:50 crc kubenswrapper[4983]: I1001 09:28:50.930530 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-c322-account-create-ttpj4"] Oct 01 09:28:51 crc kubenswrapper[4983]: I1001 09:28:51.002958 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpxn8\" (UniqueName: \"kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8\") pod \"glance-c322-account-create-ttpj4\" (UID: \"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339\") " pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:51 crc kubenswrapper[4983]: I1001 09:28:51.104357 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpxn8\" (UniqueName: \"kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8\") pod \"glance-c322-account-create-ttpj4\" (UID: \"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339\") " pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:51 crc kubenswrapper[4983]: I1001 09:28:51.126120 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpxn8\" (UniqueName: \"kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8\") pod \"glance-c322-account-create-ttpj4\" (UID: \"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339\") " pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:51 crc kubenswrapper[4983]: I1001 09:28:51.242466 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:51 crc kubenswrapper[4983]: I1001 09:28:51.643876 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-c322-account-create-ttpj4"] Oct 01 09:28:52 crc kubenswrapper[4983]: I1001 09:28:52.638087 4983 generic.go:334] "Generic (PLEG): container finished" podID="d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" containerID="4b755735c16e1f9f98f7038c8cd883ad33e338aa935bcd4c0f060c34e39d4ca6" exitCode=0 Oct 01 09:28:52 crc kubenswrapper[4983]: I1001 09:28:52.638236 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" event={"ID":"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339","Type":"ContainerDied","Data":"4b755735c16e1f9f98f7038c8cd883ad33e338aa935bcd4c0f060c34e39d4ca6"} Oct 01 09:28:52 crc kubenswrapper[4983]: I1001 09:28:52.638410 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" event={"ID":"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339","Type":"ContainerStarted","Data":"bac313c20f08bd81a829401808cf61f4243983e7bc0dd06f19c05aa10c7da862"} Oct 01 09:28:53 crc kubenswrapper[4983]: I1001 09:28:53.895537 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.041621 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpxn8\" (UniqueName: \"kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8\") pod \"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339\" (UID: \"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339\") " Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.048042 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8" (OuterVolumeSpecName: "kube-api-access-hpxn8") pod "d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" (UID: "d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339"). InnerVolumeSpecName "kube-api-access-hpxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.143890 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpxn8\" (UniqueName: \"kubernetes.io/projected/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339-kube-api-access-hpxn8\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.653713 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" event={"ID":"d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339","Type":"ContainerDied","Data":"bac313c20f08bd81a829401808cf61f4243983e7bc0dd06f19c05aa10c7da862"} Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.653753 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bac313c20f08bd81a829401808cf61f4243983e7bc0dd06f19c05aa10c7da862" Oct 01 09:28:54 crc kubenswrapper[4983]: I1001 09:28:54.653770 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c322-account-create-ttpj4" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.961159 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-5khjf"] Oct 01 09:28:55 crc kubenswrapper[4983]: E1001 09:28:55.961826 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" containerName="mariadb-account-create" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.961848 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" containerName="mariadb-account-create" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.962013 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" containerName="mariadb-account-create" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.962612 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.964949 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.966292 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-f78m2" Oct 01 09:28:55 crc kubenswrapper[4983]: I1001 09:28:55.979331 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5khjf"] Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.070119 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.070225 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.070394 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtl97\" (UniqueName: \"kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.172237 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.172308 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtl97\" (UniqueName: \"kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.172365 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.179595 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.179635 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.190570 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtl97\" (UniqueName: \"kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97\") pod \"glance-db-sync-5khjf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.279550 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:28:56 crc kubenswrapper[4983]: I1001 09:28:56.672546 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5khjf"] Oct 01 09:28:56 crc kubenswrapper[4983]: W1001 09:28:56.677339 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf774b9b8_52f2_4c3b_a6f7_16aafc4c3bdf.slice/crio-3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5 WatchSource:0}: Error finding container 3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5: Status 404 returned error can't find the container with id 3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5 Oct 01 09:28:57 crc kubenswrapper[4983]: I1001 09:28:57.684953 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5khjf" event={"ID":"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf","Type":"ContainerStarted","Data":"e7c59ebf221c2973d554934f4135cdee58ba2881abba30b0ac0df420eff0f9ca"} Oct 01 09:28:57 crc kubenswrapper[4983]: I1001 09:28:57.685323 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5khjf" event={"ID":"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf","Type":"ContainerStarted","Data":"3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5"} Oct 01 09:28:57 crc kubenswrapper[4983]: I1001 09:28:57.703628 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-5khjf" podStartSLOduration=2.703610525 podStartE2EDuration="2.703610525s" podCreationTimestamp="2025-10-01 09:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:28:57.699765161 +0000 UTC m=+1165.688993948" watchObservedRunningTime="2025-10-01 09:28:57.703610525 +0000 UTC m=+1165.692839322" Oct 01 09:29:00 crc kubenswrapper[4983]: I1001 09:29:00.707453 4983 generic.go:334] "Generic (PLEG): container finished" podID="f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" containerID="e7c59ebf221c2973d554934f4135cdee58ba2881abba30b0ac0df420eff0f9ca" exitCode=0 Oct 01 09:29:00 crc kubenswrapper[4983]: I1001 09:29:00.707553 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5khjf" event={"ID":"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf","Type":"ContainerDied","Data":"e7c59ebf221c2973d554934f4135cdee58ba2881abba30b0ac0df420eff0f9ca"} Oct 01 09:29:01 crc kubenswrapper[4983]: I1001 09:29:01.974068 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.051015 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtl97\" (UniqueName: \"kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97\") pod \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.051108 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data\") pod \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.051155 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data\") pod \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\" (UID: \"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf\") " Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.056476 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" (UID: "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.056480 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97" (OuterVolumeSpecName: "kube-api-access-qtl97") pod "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" (UID: "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf"). InnerVolumeSpecName "kube-api-access-qtl97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.086407 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data" (OuterVolumeSpecName: "config-data") pod "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" (UID: "f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.153504 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtl97\" (UniqueName: \"kubernetes.io/projected/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-kube-api-access-qtl97\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.153962 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.154720 4983 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.724080 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-5khjf" Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.724486 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-5khjf" event={"ID":"f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf","Type":"ContainerDied","Data":"3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5"} Oct 01 09:29:02 crc kubenswrapper[4983]: I1001 09:29:02.724526 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bd539af94d4525a2142085e2b202f2a5eeb1c3e03b9b6bbea39a7c37be5c5f5" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.828080 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 01 09:29:03 crc kubenswrapper[4983]: E1001 09:29:03.828316 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" containerName="glance-db-sync" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.828328 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" containerName="glance-db-sync" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.828504 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" containerName="glance-db-sync" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.829537 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.831181 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.831431 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-f78m2" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.832008 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.845585 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976516 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976605 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-config-data\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976633 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976702 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-sys\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976742 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976773 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976864 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.976899 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdvhl\" (UniqueName: \"kubernetes.io/projected/28b127a4-30d1-44a6-ac91-d3bb63546626-kube-api-access-bdvhl\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977034 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977410 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-logs\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977554 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-dev\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977794 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-scripts\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977896 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:03 crc kubenswrapper[4983]: I1001 09:29:03.977941 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.063897 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.065180 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.068091 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079404 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-scripts\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079457 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079480 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079538 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-config-data\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079565 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079594 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-sys\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079604 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079666 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079666 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079617 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079733 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079773 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079798 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdvhl\" (UniqueName: \"kubernetes.io/projected/28b127a4-30d1-44a6-ac91-d3bb63546626-kube-api-access-bdvhl\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079846 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079875 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-logs\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079903 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-dev\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.079977 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-dev\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.080316 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.080660 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.080791 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.080967 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.081021 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-sys\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.081195 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28b127a4-30d1-44a6-ac91-d3bb63546626-logs\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.080610 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/28b127a4-30d1-44a6-ac91-d3bb63546626-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.081346 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.085483 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-scripts\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.096856 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28b127a4-30d1-44a6-ac91-d3bb63546626-config-data\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.099188 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdvhl\" (UniqueName: \"kubernetes.io/projected/28b127a4-30d1-44a6-ac91-d3bb63546626-kube-api-access-bdvhl\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.105201 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.106519 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"28b127a4-30d1-44a6-ac91-d3bb63546626\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.155232 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181408 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181444 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181461 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181499 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np7r5\" (UniqueName: \"kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181526 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181559 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181630 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181648 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181667 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181687 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181714 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181729 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.181751 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282783 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282850 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282883 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282919 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282957 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np7r5\" (UniqueName: \"kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.282980 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283007 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283033 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283050 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283072 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283098 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283133 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283158 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283187 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283305 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283752 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.283939 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.284017 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286246 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286305 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286329 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286353 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286354 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286503 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.286630 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.296392 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.302491 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.313986 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.318324 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.339017 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np7r5\" (UniqueName: \"kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5\") pod \"glance-default-internal-api-0\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.385669 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.392268 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.741733 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"28b127a4-30d1-44a6-ac91-d3bb63546626","Type":"ContainerStarted","Data":"aa4bbcffbb8e1569d8dab5a712c74f77060812d46f2585c1a209b9143b55f7e3"} Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.742423 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"28b127a4-30d1-44a6-ac91-d3bb63546626","Type":"ContainerStarted","Data":"c86a7eb989a8a29ad880102e132a59b543f43aa450e950aa4fa94376e92157a4"} Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.742437 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"28b127a4-30d1-44a6-ac91-d3bb63546626","Type":"ContainerStarted","Data":"f1a7fac00bb32bd0743c463c4116904a08c78dda3c1b5842782159f4ad417a31"} Oct 01 09:29:04 crc kubenswrapper[4983]: I1001 09:29:04.830696 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:04 crc kubenswrapper[4983]: W1001 09:29:04.835451 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf830a1b_a8a6_40b6_8249_adfff624d020.slice/crio-6a8ce3addb5ec9aaff264d1396c2f032e964fc27fcf5ade73060d9d5e17d0aa1 WatchSource:0}: Error finding container 6a8ce3addb5ec9aaff264d1396c2f032e964fc27fcf5ade73060d9d5e17d0aa1: Status 404 returned error can't find the container with id 6a8ce3addb5ec9aaff264d1396c2f032e964fc27fcf5ade73060d9d5e17d0aa1 Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.265128 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.750384 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerStarted","Data":"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565"} Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.750429 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerStarted","Data":"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0"} Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.750441 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerStarted","Data":"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed"} Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.750450 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerStarted","Data":"6a8ce3addb5ec9aaff264d1396c2f032e964fc27fcf5ade73060d9d5e17d0aa1"} Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.750578 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-log" containerID="cri-o://fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" gracePeriod=30 Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.751079 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-api" containerID="cri-o://0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" gracePeriod=30 Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.751126 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-httpd" containerID="cri-o://42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" gracePeriod=30 Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.755627 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"28b127a4-30d1-44a6-ac91-d3bb63546626","Type":"ContainerStarted","Data":"5c379b4264d089e955e86cb8db4b54cc7c74d372b9e9fb238175ce781a66cee5"} Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.784658 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.784637839 podStartE2EDuration="2.784637839s" podCreationTimestamp="2025-10-01 09:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:29:05.779560837 +0000 UTC m=+1173.768789654" watchObservedRunningTime="2025-10-01 09:29:05.784637839 +0000 UTC m=+1173.773866636" Oct 01 09:29:05 crc kubenswrapper[4983]: I1001 09:29:05.821387 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.821363476 podStartE2EDuration="2.821363476s" podCreationTimestamp="2025-10-01 09:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:29:05.809216273 +0000 UTC m=+1173.798445070" watchObservedRunningTime="2025-10-01 09:29:05.821363476 +0000 UTC m=+1173.810592273" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.119445 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.211859 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.211945 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.211994 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.211985 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212054 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212089 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212113 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212145 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212164 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run" (OuterVolumeSpecName: "run") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212190 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np7r5\" (UniqueName: \"kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212202 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys" (OuterVolumeSpecName: "sys") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212238 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212266 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212278 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs" (OuterVolumeSpecName: "logs") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212323 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212325 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212323 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212345 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212365 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev" (OuterVolumeSpecName: "dev") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212387 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212535 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212597 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212627 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi\") pod \"cf830a1b-a8a6-40b6-8249-adfff624d020\" (UID: \"cf830a1b-a8a6-40b6-8249-adfff624d020\") " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212962 4983 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212984 4983 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-sys\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.212995 4983 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213004 4983 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213014 4983 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-dev\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213023 4983 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213034 4983 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213043 4983 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cf830a1b-a8a6-40b6-8249-adfff624d020-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.213053 4983 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf830a1b-a8a6-40b6-8249-adfff624d020-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.217736 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.217837 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts" (OuterVolumeSpecName: "scripts") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.217840 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.217955 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5" (OuterVolumeSpecName: "kube-api-access-np7r5") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "kube-api-access-np7r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.284728 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data" (OuterVolumeSpecName: "config-data") pod "cf830a1b-a8a6-40b6-8249-adfff624d020" (UID: "cf830a1b-a8a6-40b6-8249-adfff624d020"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.314061 4983 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.314116 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.314126 4983 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf830a1b-a8a6-40b6-8249-adfff624d020-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.314136 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np7r5\" (UniqueName: \"kubernetes.io/projected/cf830a1b-a8a6-40b6-8249-adfff624d020-kube-api-access-np7r5\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.314151 4983 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.327306 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.332579 4983 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.415969 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.416008 4983 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763753 4983 generic.go:334] "Generic (PLEG): container finished" podID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" exitCode=143 Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763783 4983 generic.go:334] "Generic (PLEG): container finished" podID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" exitCode=143 Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763790 4983 generic.go:334] "Generic (PLEG): container finished" podID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" exitCode=143 Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763827 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763897 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerDied","Data":"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565"} Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763922 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerDied","Data":"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0"} Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763932 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerDied","Data":"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed"} Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763942 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"cf830a1b-a8a6-40b6-8249-adfff624d020","Type":"ContainerDied","Data":"6a8ce3addb5ec9aaff264d1396c2f032e964fc27fcf5ade73060d9d5e17d0aa1"} Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.763956 4983 scope.go:117] "RemoveContainer" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.783749 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.789778 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.792730 4983 scope.go:117] "RemoveContainer" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812430 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.812713 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-log" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812728 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-log" Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.812747 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-httpd" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812754 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-httpd" Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.812764 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-api" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812769 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-api" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812913 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-httpd" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812930 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-log" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.812945 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" containerName="glance-api" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.813853 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.814829 4983 scope.go:117] "RemoveContainer" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.816563 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.843771 4983 scope.go:117] "RemoveContainer" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.844341 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": container with ID starting with 0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565 not found: ID does not exist" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.844384 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.844398 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565"} err="failed to get container status \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": rpc error: code = NotFound desc = could not find container \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": container with ID starting with 0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.844434 4983 scope.go:117] "RemoveContainer" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.844862 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": container with ID starting with 42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0 not found: ID does not exist" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.844893 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0"} err="failed to get container status \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": rpc error: code = NotFound desc = could not find container \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": container with ID starting with 42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.844918 4983 scope.go:117] "RemoveContainer" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" Oct 01 09:29:06 crc kubenswrapper[4983]: E1001 09:29:06.845396 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": container with ID starting with fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed not found: ID does not exist" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.845417 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed"} err="failed to get container status \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": rpc error: code = NotFound desc = could not find container \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": container with ID starting with fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.845437 4983 scope.go:117] "RemoveContainer" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.845676 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565"} err="failed to get container status \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": rpc error: code = NotFound desc = could not find container \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": container with ID starting with 0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.845720 4983 scope.go:117] "RemoveContainer" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.846751 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0"} err="failed to get container status \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": rpc error: code = NotFound desc = could not find container \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": container with ID starting with 42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.846781 4983 scope.go:117] "RemoveContainer" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847049 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed"} err="failed to get container status \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": rpc error: code = NotFound desc = could not find container \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": container with ID starting with fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847068 4983 scope.go:117] "RemoveContainer" containerID="0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847267 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565"} err="failed to get container status \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": rpc error: code = NotFound desc = could not find container \"0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565\": container with ID starting with 0c7d469bd8efb4cf5e8bc7dd6dcd38727759b59c71e5ac83244f3931dd900565 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847298 4983 scope.go:117] "RemoveContainer" containerID="42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847504 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0"} err="failed to get container status \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": rpc error: code = NotFound desc = could not find container \"42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0\": container with ID starting with 42f3ddcaf347b613e4edc865e73e38b8f0eeb7763796b06f3a435c478e6359d0 not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847529 4983 scope.go:117] "RemoveContainer" containerID="fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.847716 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed"} err="failed to get container status \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": rpc error: code = NotFound desc = could not find container \"fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed\": container with ID starting with fa22cc7e5978bcece44d697cee7e5a347e2f0ee9e48d9cdf9c5e717269217eed not found: ID does not exist" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923618 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923824 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923859 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923885 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbxt\" (UniqueName: \"kubernetes.io/projected/3dc1ff50-c96e-4360-a2c4-954f495f107a-kube-api-access-6sbxt\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923907 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923937 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923964 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.923982 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924003 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924106 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924185 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924229 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924303 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:06 crc kubenswrapper[4983]: I1001 09:29:06.924346 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025582 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025673 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025710 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbxt\" (UniqueName: \"kubernetes.io/projected/3dc1ff50-c96e-4360-a2c4-954f495f107a-kube-api-access-6sbxt\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025735 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025769 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025784 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025801 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025863 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025890 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025911 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025932 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025928 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025964 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025986 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.025989 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026024 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026074 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026263 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026532 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026552 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026587 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-logs\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026608 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ff50-c96e-4360-a2c4-954f495f107a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026644 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-sys\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.026653 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.028448 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3dc1ff50-c96e-4360-a2c4-954f495f107a-dev\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.036899 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.038010 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ff50-c96e-4360-a2c4-954f495f107a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.046542 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.047940 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbxt\" (UniqueName: \"kubernetes.io/projected/3dc1ff50-c96e-4360-a2c4-954f495f107a-kube-api-access-6sbxt\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.052105 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3dc1ff50-c96e-4360-a2c4-954f495f107a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.139946 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.560943 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.780274 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3dc1ff50-c96e-4360-a2c4-954f495f107a","Type":"ContainerStarted","Data":"52d27c49dbc2f79111b512c1c7fd13493f21598f85142117fde3e865b0729648"} Oct 01 09:29:07 crc kubenswrapper[4983]: I1001 09:29:07.780674 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3dc1ff50-c96e-4360-a2c4-954f495f107a","Type":"ContainerStarted","Data":"1cea6b96f3dd7fab4580e77658f40afeab6d9d50c69fc9463dc16ecd4511884a"} Oct 01 09:29:08 crc kubenswrapper[4983]: I1001 09:29:08.729208 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf830a1b-a8a6-40b6-8249-adfff624d020" path="/var/lib/kubelet/pods/cf830a1b-a8a6-40b6-8249-adfff624d020/volumes" Oct 01 09:29:08 crc kubenswrapper[4983]: I1001 09:29:08.815187 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3dc1ff50-c96e-4360-a2c4-954f495f107a","Type":"ContainerStarted","Data":"2b785a54ba1c7c1369b46227ac4d2b7c2a3b95e876cb628325c8594e8e7340b8"} Oct 01 09:29:08 crc kubenswrapper[4983]: I1001 09:29:08.815237 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"3dc1ff50-c96e-4360-a2c4-954f495f107a","Type":"ContainerStarted","Data":"f1f88a8de2378e63e3b7b0a83c2ed6a9c4631d957fe6129eab5610310e11aacc"} Oct 01 09:29:08 crc kubenswrapper[4983]: I1001 09:29:08.846027 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.8460042420000002 podStartE2EDuration="2.846004242s" podCreationTimestamp="2025-10-01 09:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:29:08.843154309 +0000 UTC m=+1176.832383116" watchObservedRunningTime="2025-10-01 09:29:08.846004242 +0000 UTC m=+1176.835233039" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.156052 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.156510 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.156540 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.185142 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.188412 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.198082 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.864698 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.864734 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.864748 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.879705 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.880145 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:14 crc kubenswrapper[4983]: I1001 09:29:14.884534 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.141539 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.141952 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.141966 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.162407 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.166797 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.188327 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.886575 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.886632 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.886649 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.898890 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.899353 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:17 crc kubenswrapper[4983]: I1001 09:29:17.901176 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 01 09:29:31 crc kubenswrapper[4983]: I1001 09:29:31.960852 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:29:31 crc kubenswrapper[4983]: I1001 09:29:31.961431 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.168975 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.170798 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.180465 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.181489 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.189016 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.196398 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.239178 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.240282 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.246181 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.246196 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.250659 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w"] Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.265896 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2mj4\" (UniqueName: \"kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.265965 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.266007 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.266030 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.266070 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.266105 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd69r\" (UniqueName: \"kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.287882 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.287908 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.367989 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd69r\" (UniqueName: \"kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368077 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368131 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368164 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2mj4\" (UniqueName: \"kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368281 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfllk\" (UniqueName: \"kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368407 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.368462 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.374242 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.374521 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.391842 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd69r\" (UniqueName: \"kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r\") pod \"glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.392652 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2mj4\" (UniqueName: \"kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.469748 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.470147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.470182 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfllk\" (UniqueName: \"kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.471479 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.473727 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.491522 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfllk\" (UniqueName: \"kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk\") pod \"collect-profiles-29321850-hj84w\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.493918 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.505786 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.561964 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:00 crc kubenswrapper[4983]: I1001 09:30:00.944369 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb"] Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.001575 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8"] Oct 01 09:30:01 crc kubenswrapper[4983]: W1001 09:30:01.006770 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c9195c3_eb97_4802_9026_f496fd13e680.slice/crio-5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668 WatchSource:0}: Error finding container 5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668: Status 404 returned error can't find the container with id 5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668 Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.063887 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w"] Oct 01 09:30:01 crc kubenswrapper[4983]: W1001 09:30:01.072515 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71d9d0a7_9e1b_4912_9f77_f6f50c0fe6d1.slice/crio-e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04 WatchSource:0}: Error finding container e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04: Status 404 returned error can't find the container with id e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04 Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.203900 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" event={"ID":"9c9195c3-eb97-4802-9026-f496fd13e680","Type":"ContainerStarted","Data":"5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668"} Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.207336 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" event={"ID":"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1","Type":"ContainerStarted","Data":"e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04"} Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.210930 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" event={"ID":"eebb3506-fc09-4c67-bd25-df8ee0cc59ad","Type":"ContainerStarted","Data":"e10bd0030f545ba5c08ba3faa5ee17732325396c67d231856bbd8028f76bae14"} Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.961017 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:30:01 crc kubenswrapper[4983]: I1001 09:30:01.962081 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.222104 4983 generic.go:334] "Generic (PLEG): container finished" podID="71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" containerID="c585948704f11ae4858eea84b1ef0d884f733713e4b5da2856a65fd9ca99d7f4" exitCode=0 Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.222214 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" event={"ID":"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1","Type":"ContainerDied","Data":"c585948704f11ae4858eea84b1ef0d884f733713e4b5da2856a65fd9ca99d7f4"} Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.229331 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" event={"ID":"eebb3506-fc09-4c67-bd25-df8ee0cc59ad","Type":"ContainerStarted","Data":"ea7a6d715ffac2c9b9ce75643ed9b7b21619554ef5ab16bc2be896aa0ba39e11"} Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.231347 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" event={"ID":"9c9195c3-eb97-4802-9026-f496fd13e680","Type":"ContainerStarted","Data":"bb767364c7e68d2c186bb0d8b367f07795767e3849bf04665354210e99d3a261"} Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.264858 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" podStartSLOduration=2.264837844 podStartE2EDuration="2.264837844s" podCreationTimestamp="2025-10-01 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:30:02.261410076 +0000 UTC m=+1230.250638893" watchObservedRunningTime="2025-10-01 09:30:02.264837844 +0000 UTC m=+1230.254066641" Oct 01 09:30:02 crc kubenswrapper[4983]: I1001 09:30:02.289740 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" podStartSLOduration=2.289719287 podStartE2EDuration="2.289719287s" podCreationTimestamp="2025-10-01 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:30:02.288083854 +0000 UTC m=+1230.277312651" watchObservedRunningTime="2025-10-01 09:30:02.289719287 +0000 UTC m=+1230.278948084" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.239939 4983 generic.go:334] "Generic (PLEG): container finished" podID="eebb3506-fc09-4c67-bd25-df8ee0cc59ad" containerID="ea7a6d715ffac2c9b9ce75643ed9b7b21619554ef5ab16bc2be896aa0ba39e11" exitCode=0 Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.240068 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" event={"ID":"eebb3506-fc09-4c67-bd25-df8ee0cc59ad","Type":"ContainerDied","Data":"ea7a6d715ffac2c9b9ce75643ed9b7b21619554ef5ab16bc2be896aa0ba39e11"} Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.241776 4983 generic.go:334] "Generic (PLEG): container finished" podID="9c9195c3-eb97-4802-9026-f496fd13e680" containerID="bb767364c7e68d2c186bb0d8b367f07795767e3849bf04665354210e99d3a261" exitCode=0 Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.241873 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" event={"ID":"9c9195c3-eb97-4802-9026-f496fd13e680","Type":"ContainerDied","Data":"bb767364c7e68d2c186bb0d8b367f07795767e3849bf04665354210e99d3a261"} Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.524553 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.623453 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume\") pod \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.623525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume\") pod \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.623668 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfllk\" (UniqueName: \"kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk\") pod \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\" (UID: \"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1\") " Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.624912 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" (UID: "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.631384 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" (UID: "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.631581 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk" (OuterVolumeSpecName: "kube-api-access-jfllk") pod "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" (UID: "71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1"). InnerVolumeSpecName "kube-api-access-jfllk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.726696 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfllk\" (UniqueName: \"kubernetes.io/projected/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-kube-api-access-jfllk\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.726745 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:03 crc kubenswrapper[4983]: I1001 09:30:03.726758 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.251384 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" event={"ID":"71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1","Type":"ContainerDied","Data":"e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04"} Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.251906 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c62c19b747ac8b15bd2a393e0e032a1a9e14c9f9f6cb1ce8235c2da2829f04" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.251428 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-hj84w" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.562534 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.569427 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643094 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9c9195c3-eb97-4802-9026-f496fd13e680\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643201 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd69r\" (UniqueName: \"kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r\") pod \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643223 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2mj4\" (UniqueName: \"kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4\") pod \"9c9195c3-eb97-4802-9026-f496fd13e680\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643242 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643271 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data\") pod \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\" (UID: \"eebb3506-fc09-4c67-bd25-df8ee0cc59ad\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.643369 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data\") pod \"9c9195c3-eb97-4802-9026-f496fd13e680\" (UID: \"9c9195c3-eb97-4802-9026-f496fd13e680\") " Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.648455 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "9c9195c3-eb97-4802-9026-f496fd13e680" (UID: "9c9195c3-eb97-4802-9026-f496fd13e680"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.649317 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "eebb3506-fc09-4c67-bd25-df8ee0cc59ad" (UID: "eebb3506-fc09-4c67-bd25-df8ee0cc59ad"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.649415 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "eebb3506-fc09-4c67-bd25-df8ee0cc59ad" (UID: "eebb3506-fc09-4c67-bd25-df8ee0cc59ad"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.649506 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4" (OuterVolumeSpecName: "kube-api-access-m2mj4") pod "9c9195c3-eb97-4802-9026-f496fd13e680" (UID: "9c9195c3-eb97-4802-9026-f496fd13e680"). InnerVolumeSpecName "kube-api-access-m2mj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.650614 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "9c9195c3-eb97-4802-9026-f496fd13e680" (UID: "9c9195c3-eb97-4802-9026-f496fd13e680"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.651069 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r" (OuterVolumeSpecName: "kube-api-access-pd69r") pod "eebb3506-fc09-4c67-bd25-df8ee0cc59ad" (UID: "eebb3506-fc09-4c67-bd25-df8ee0cc59ad"). InnerVolumeSpecName "kube-api-access-pd69r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.744189 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd69r\" (UniqueName: \"kubernetes.io/projected/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-kube-api-access-pd69r\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.744665 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2mj4\" (UniqueName: \"kubernetes.io/projected/9c9195c3-eb97-4802-9026-f496fd13e680-kube-api-access-m2mj4\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.744726 4983 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eebb3506-fc09-4c67-bd25-df8ee0cc59ad-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:04 crc kubenswrapper[4983]: I1001 09:30:04.744794 4983 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/9c9195c3-eb97-4802-9026-f496fd13e680-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.258816 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.258815 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8" event={"ID":"9c9195c3-eb97-4802-9026-f496fd13e680","Type":"ContainerDied","Data":"5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668"} Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.258937 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d2c715ba80b6698cbf32572cafc96c8e7b130e769b487d3d850382c77ee5668" Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.260580 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" event={"ID":"eebb3506-fc09-4c67-bd25-df8ee0cc59ad","Type":"ContainerDied","Data":"e10bd0030f545ba5c08ba3faa5ee17732325396c67d231856bbd8028f76bae14"} Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.260599 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb" Oct 01 09:30:05 crc kubenswrapper[4983]: I1001 09:30:05.260614 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e10bd0030f545ba5c08ba3faa5ee17732325396c67d231856bbd8028f76bae14" Oct 01 09:30:31 crc kubenswrapper[4983]: I1001 09:30:31.961281 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:30:31 crc kubenswrapper[4983]: I1001 09:30:31.962022 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:30:31 crc kubenswrapper[4983]: I1001 09:30:31.962087 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:30:31 crc kubenswrapper[4983]: I1001 09:30:31.962858 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:30:31 crc kubenswrapper[4983]: I1001 09:30:31.962959 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50" gracePeriod=600 Oct 01 09:30:32 crc kubenswrapper[4983]: I1001 09:30:32.506181 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50" exitCode=0 Oct 01 09:30:32 crc kubenswrapper[4983]: I1001 09:30:32.506452 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50"} Oct 01 09:30:32 crc kubenswrapper[4983]: I1001 09:30:32.506761 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7"} Oct 01 09:30:32 crc kubenswrapper[4983]: I1001 09:30:32.506821 4983 scope.go:117] "RemoveContainer" containerID="3fbd4985585bf219cf32170591b894da2c23a255b07d456ee1d008e3d414c025" Oct 01 09:32:33 crc kubenswrapper[4983]: I1001 09:32:33.330666 4983 scope.go:117] "RemoveContainer" containerID="750db2a684d3f813b2811ae3e5aeb15fdcbc0abbd62dbaf3f34669faa4561f38" Oct 01 09:32:33 crc kubenswrapper[4983]: I1001 09:32:33.357520 4983 scope.go:117] "RemoveContainer" containerID="be88385fa74c8a6e25ecd1d2ac544f0bd5e40dbe2c2d2541cc0080f0bef6f017" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.158427 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:34 crc kubenswrapper[4983]: E1001 09:32:34.159307 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebb3506-fc09-4c67-bd25-df8ee0cc59ad" containerName="glance-cache-glance-default-external-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.159406 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebb3506-fc09-4c67-bd25-df8ee0cc59ad" containerName="glance-cache-glance-default-external-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: E1001 09:32:34.159523 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9195c3-eb97-4802-9026-f496fd13e680" containerName="glance-cache-glance-default-internal-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.159608 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9195c3-eb97-4802-9026-f496fd13e680" containerName="glance-cache-glance-default-internal-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: E1001 09:32:34.159713 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" containerName="collect-profiles" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.159788 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" containerName="collect-profiles" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.160031 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c9195c3-eb97-4802-9026-f496fd13e680" containerName="glance-cache-glance-default-internal-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.160130 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebb3506-fc09-4c67-bd25-df8ee0cc59ad" containerName="glance-cache-glance-default-external-api-0-cleaner" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.160211 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="71d9d0a7-9e1b-4912-9f77-f6f50c0fe6d1" containerName="collect-profiles" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.161508 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.172235 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.209770 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.209975 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9flt2\" (UniqueName: \"kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.210027 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.311966 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9flt2\" (UniqueName: \"kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.312040 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.312127 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.312690 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.312776 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.334966 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9flt2\" (UniqueName: \"kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2\") pod \"community-operators-7zrtr\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.487984 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:34 crc kubenswrapper[4983]: I1001 09:32:34.927490 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:35 crc kubenswrapper[4983]: I1001 09:32:35.494896 4983 generic.go:334] "Generic (PLEG): container finished" podID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerID="28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77" exitCode=0 Oct 01 09:32:35 crc kubenswrapper[4983]: I1001 09:32:35.494979 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerDied","Data":"28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77"} Oct 01 09:32:35 crc kubenswrapper[4983]: I1001 09:32:35.495184 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerStarted","Data":"95c6eb2e494b6ff97b2fe5387d291b5a6d82e855879c099f832449c0d86fee3a"} Oct 01 09:32:35 crc kubenswrapper[4983]: I1001 09:32:35.497665 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:32:37 crc kubenswrapper[4983]: I1001 09:32:37.516145 4983 generic.go:334] "Generic (PLEG): container finished" podID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerID="d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d" exitCode=0 Oct 01 09:32:37 crc kubenswrapper[4983]: I1001 09:32:37.516243 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerDied","Data":"d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d"} Oct 01 09:32:38 crc kubenswrapper[4983]: I1001 09:32:38.527334 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerStarted","Data":"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b"} Oct 01 09:32:38 crc kubenswrapper[4983]: I1001 09:32:38.552505 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7zrtr" podStartSLOduration=2.101501915 podStartE2EDuration="4.55248137s" podCreationTimestamp="2025-10-01 09:32:34 +0000 UTC" firstStartedPulling="2025-10-01 09:32:35.49738042 +0000 UTC m=+1383.486609218" lastFinishedPulling="2025-10-01 09:32:37.948359876 +0000 UTC m=+1385.937588673" observedRunningTime="2025-10-01 09:32:38.545050644 +0000 UTC m=+1386.534279441" watchObservedRunningTime="2025-10-01 09:32:38.55248137 +0000 UTC m=+1386.541710167" Oct 01 09:32:44 crc kubenswrapper[4983]: I1001 09:32:44.489202 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:44 crc kubenswrapper[4983]: I1001 09:32:44.489971 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:44 crc kubenswrapper[4983]: I1001 09:32:44.526888 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:44 crc kubenswrapper[4983]: I1001 09:32:44.627528 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:44 crc kubenswrapper[4983]: I1001 09:32:44.763835 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:46 crc kubenswrapper[4983]: I1001 09:32:46.608237 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7zrtr" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="registry-server" containerID="cri-o://14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b" gracePeriod=2 Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.047470 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.214413 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9flt2\" (UniqueName: \"kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2\") pod \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.214646 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities\") pod \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.214757 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content\") pod \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\" (UID: \"f56d05e2-8c05-48bd-a46e-00a808a7e72f\") " Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.215630 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities" (OuterVolumeSpecName: "utilities") pod "f56d05e2-8c05-48bd-a46e-00a808a7e72f" (UID: "f56d05e2-8c05-48bd-a46e-00a808a7e72f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.220200 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2" (OuterVolumeSpecName: "kube-api-access-9flt2") pod "f56d05e2-8c05-48bd-a46e-00a808a7e72f" (UID: "f56d05e2-8c05-48bd-a46e-00a808a7e72f"). InnerVolumeSpecName "kube-api-access-9flt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.316821 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.316865 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9flt2\" (UniqueName: \"kubernetes.io/projected/f56d05e2-8c05-48bd-a46e-00a808a7e72f-kube-api-access-9flt2\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.620109 4983 generic.go:334] "Generic (PLEG): container finished" podID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerID="14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b" exitCode=0 Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.620219 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerDied","Data":"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b"} Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.620293 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zrtr" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.620323 4983 scope.go:117] "RemoveContainer" containerID="14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.620307 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zrtr" event={"ID":"f56d05e2-8c05-48bd-a46e-00a808a7e72f","Type":"ContainerDied","Data":"95c6eb2e494b6ff97b2fe5387d291b5a6d82e855879c099f832449c0d86fee3a"} Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.647911 4983 scope.go:117] "RemoveContainer" containerID="d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.676125 4983 scope.go:117] "RemoveContainer" containerID="28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.720638 4983 scope.go:117] "RemoveContainer" containerID="14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b" Oct 01 09:32:47 crc kubenswrapper[4983]: E1001 09:32:47.721380 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b\": container with ID starting with 14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b not found: ID does not exist" containerID="14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.721440 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b"} err="failed to get container status \"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b\": rpc error: code = NotFound desc = could not find container \"14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b\": container with ID starting with 14aaa47b728559ad9e6688973a49348444af95f39d598d1569f1bc665c88d89b not found: ID does not exist" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.721474 4983 scope.go:117] "RemoveContainer" containerID="d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d" Oct 01 09:32:47 crc kubenswrapper[4983]: E1001 09:32:47.722142 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d\": container with ID starting with d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d not found: ID does not exist" containerID="d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.722221 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d"} err="failed to get container status \"d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d\": rpc error: code = NotFound desc = could not find container \"d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d\": container with ID starting with d4a5ed2a84399e542bd72a5c84190f7d40f728df1e2db1632179805733d8990d not found: ID does not exist" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.722266 4983 scope.go:117] "RemoveContainer" containerID="28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77" Oct 01 09:32:47 crc kubenswrapper[4983]: E1001 09:32:47.722786 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77\": container with ID starting with 28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77 not found: ID does not exist" containerID="28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.722986 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77"} err="failed to get container status \"28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77\": rpc error: code = NotFound desc = could not find container \"28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77\": container with ID starting with 28c7f7f86a8ae38da6078830ad6b35a6a81c72562955510e4c2f02e01c67ad77 not found: ID does not exist" Oct 01 09:32:47 crc kubenswrapper[4983]: I1001 09:32:47.934908 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f56d05e2-8c05-48bd-a46e-00a808a7e72f" (UID: "f56d05e2-8c05-48bd-a46e-00a808a7e72f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:32:48 crc kubenswrapper[4983]: I1001 09:32:48.028581 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56d05e2-8c05-48bd-a46e-00a808a7e72f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:48 crc kubenswrapper[4983]: I1001 09:32:48.251609 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:48 crc kubenswrapper[4983]: I1001 09:32:48.258063 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7zrtr"] Oct 01 09:32:48 crc kubenswrapper[4983]: I1001 09:32:48.722369 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" path="/var/lib/kubelet/pods/f56d05e2-8c05-48bd-a46e-00a808a7e72f/volumes" Oct 01 09:33:01 crc kubenswrapper[4983]: I1001 09:33:01.961542 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:33:01 crc kubenswrapper[4983]: I1001 09:33:01.962347 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.022263 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:15 crc kubenswrapper[4983]: E1001 09:33:15.023068 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="extract-content" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.023081 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="extract-content" Oct 01 09:33:15 crc kubenswrapper[4983]: E1001 09:33:15.023105 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="registry-server" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.023111 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="registry-server" Oct 01 09:33:15 crc kubenswrapper[4983]: E1001 09:33:15.023128 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="extract-utilities" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.023135 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="extract-utilities" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.023266 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56d05e2-8c05-48bd-a46e-00a808a7e72f" containerName="registry-server" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.024197 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.035230 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.147569 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.147632 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.147742 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt6ls\" (UniqueName: \"kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.249161 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.249212 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.249272 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt6ls\" (UniqueName: \"kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.250072 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.250125 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.269890 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt6ls\" (UniqueName: \"kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls\") pod \"redhat-marketplace-8ppn2\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.343104 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.741434 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:15 crc kubenswrapper[4983]: I1001 09:33:15.861387 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerStarted","Data":"ac50bfe1aa0fb03dbf52a49e24647e4802f0b113b02e4d240da21119d9368160"} Oct 01 09:33:16 crc kubenswrapper[4983]: I1001 09:33:16.870331 4983 generic.go:334] "Generic (PLEG): container finished" podID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerID="192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2" exitCode=0 Oct 01 09:33:16 crc kubenswrapper[4983]: I1001 09:33:16.870385 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerDied","Data":"192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2"} Oct 01 09:33:18 crc kubenswrapper[4983]: I1001 09:33:18.887643 4983 generic.go:334] "Generic (PLEG): container finished" podID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerID="a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090" exitCode=0 Oct 01 09:33:18 crc kubenswrapper[4983]: I1001 09:33:18.887746 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerDied","Data":"a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090"} Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.811346 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.813048 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.824999 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.896504 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerStarted","Data":"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe"} Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.914121 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8ppn2" podStartSLOduration=2.459315783 podStartE2EDuration="4.914100642s" podCreationTimestamp="2025-10-01 09:33:15 +0000 UTC" firstStartedPulling="2025-10-01 09:33:16.872131634 +0000 UTC m=+1424.861360441" lastFinishedPulling="2025-10-01 09:33:19.326916503 +0000 UTC m=+1427.316145300" observedRunningTime="2025-10-01 09:33:19.914056481 +0000 UTC m=+1427.903285278" watchObservedRunningTime="2025-10-01 09:33:19.914100642 +0000 UTC m=+1427.903329439" Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.917911 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flcqj\" (UniqueName: \"kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.917974 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:19 crc kubenswrapper[4983]: I1001 09:33:19.918184 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.019789 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.019931 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flcqj\" (UniqueName: \"kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.019960 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.020457 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.020470 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.041846 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flcqj\" (UniqueName: \"kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj\") pod \"redhat-operators-zgv6c\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.132216 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:20 crc kubenswrapper[4983]: W1001 09:33:20.565529 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9009fea7_66ed_46b2_95a7_db41ace04980.slice/crio-7a9ff0d166e387244a50c6374241b287cb0ba486d740a24b2083f61d934264e6 WatchSource:0}: Error finding container 7a9ff0d166e387244a50c6374241b287cb0ba486d740a24b2083f61d934264e6: Status 404 returned error can't find the container with id 7a9ff0d166e387244a50c6374241b287cb0ba486d740a24b2083f61d934264e6 Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.565591 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.908074 4983 generic.go:334] "Generic (PLEG): container finished" podID="9009fea7-66ed-46b2-95a7-db41ace04980" containerID="ac99ce21090fe78b424c3af0dbf1e909937c1627d0eb83e10c77a2371a6f122f" exitCode=0 Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.908176 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerDied","Data":"ac99ce21090fe78b424c3af0dbf1e909937c1627d0eb83e10c77a2371a6f122f"} Oct 01 09:33:20 crc kubenswrapper[4983]: I1001 09:33:20.908232 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerStarted","Data":"7a9ff0d166e387244a50c6374241b287cb0ba486d740a24b2083f61d934264e6"} Oct 01 09:33:21 crc kubenswrapper[4983]: I1001 09:33:21.917870 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerStarted","Data":"cd6b9067b6295d6e860063c71671a4a08b7448e7c6cc1bc0f7d52ec449ebe109"} Oct 01 09:33:22 crc kubenswrapper[4983]: I1001 09:33:22.925349 4983 generic.go:334] "Generic (PLEG): container finished" podID="9009fea7-66ed-46b2-95a7-db41ace04980" containerID="cd6b9067b6295d6e860063c71671a4a08b7448e7c6cc1bc0f7d52ec449ebe109" exitCode=0 Oct 01 09:33:22 crc kubenswrapper[4983]: I1001 09:33:22.925399 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerDied","Data":"cd6b9067b6295d6e860063c71671a4a08b7448e7c6cc1bc0f7d52ec449ebe109"} Oct 01 09:33:23 crc kubenswrapper[4983]: I1001 09:33:23.933238 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerStarted","Data":"5da91bc6d9dbf6178596c28a546c5eaa0265ba65dba6ea713a1ad6ce90a7eecd"} Oct 01 09:33:23 crc kubenswrapper[4983]: I1001 09:33:23.952701 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zgv6c" podStartSLOduration=2.266917841 podStartE2EDuration="4.952685002s" podCreationTimestamp="2025-10-01 09:33:19 +0000 UTC" firstStartedPulling="2025-10-01 09:33:20.909606936 +0000 UTC m=+1428.898835733" lastFinishedPulling="2025-10-01 09:33:23.595374097 +0000 UTC m=+1431.584602894" observedRunningTime="2025-10-01 09:33:23.94796195 +0000 UTC m=+1431.937190757" watchObservedRunningTime="2025-10-01 09:33:23.952685002 +0000 UTC m=+1431.941913789" Oct 01 09:33:25 crc kubenswrapper[4983]: I1001 09:33:25.343669 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:25 crc kubenswrapper[4983]: I1001 09:33:25.344007 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:25 crc kubenswrapper[4983]: I1001 09:33:25.383435 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:25 crc kubenswrapper[4983]: I1001 09:33:25.988512 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:26 crc kubenswrapper[4983]: I1001 09:33:26.602101 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:27 crc kubenswrapper[4983]: I1001 09:33:27.962140 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8ppn2" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="registry-server" containerID="cri-o://7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe" gracePeriod=2 Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.889119 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.952210 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content\") pod \"20ab519c-9923-4ca8-a298-b0c0eea34290\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.952346 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities\") pod \"20ab519c-9923-4ca8-a298-b0c0eea34290\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.952452 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt6ls\" (UniqueName: \"kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls\") pod \"20ab519c-9923-4ca8-a298-b0c0eea34290\" (UID: \"20ab519c-9923-4ca8-a298-b0c0eea34290\") " Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.953218 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities" (OuterVolumeSpecName: "utilities") pod "20ab519c-9923-4ca8-a298-b0c0eea34290" (UID: "20ab519c-9923-4ca8-a298-b0c0eea34290"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.957273 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls" (OuterVolumeSpecName: "kube-api-access-dt6ls") pod "20ab519c-9923-4ca8-a298-b0c0eea34290" (UID: "20ab519c-9923-4ca8-a298-b0c0eea34290"). InnerVolumeSpecName "kube-api-access-dt6ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.965154 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20ab519c-9923-4ca8-a298-b0c0eea34290" (UID: "20ab519c-9923-4ca8-a298-b0c0eea34290"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.972335 4983 generic.go:334] "Generic (PLEG): container finished" podID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerID="7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe" exitCode=0 Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.972375 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerDied","Data":"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe"} Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.972400 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ppn2" event={"ID":"20ab519c-9923-4ca8-a298-b0c0eea34290","Type":"ContainerDied","Data":"ac50bfe1aa0fb03dbf52a49e24647e4802f0b113b02e4d240da21119d9368160"} Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.972417 4983 scope.go:117] "RemoveContainer" containerID="7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe" Oct 01 09:33:28 crc kubenswrapper[4983]: I1001 09:33:28.972426 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ppn2" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.005041 4983 scope.go:117] "RemoveContainer" containerID="a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.016208 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.027478 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ppn2"] Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.045701 4983 scope.go:117] "RemoveContainer" containerID="192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.055224 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.055292 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ab519c-9923-4ca8-a298-b0c0eea34290-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.055323 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt6ls\" (UniqueName: \"kubernetes.io/projected/20ab519c-9923-4ca8-a298-b0c0eea34290-kube-api-access-dt6ls\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.063226 4983 scope.go:117] "RemoveContainer" containerID="7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe" Oct 01 09:33:29 crc kubenswrapper[4983]: E1001 09:33:29.064169 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe\": container with ID starting with 7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe not found: ID does not exist" containerID="7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.064212 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe"} err="failed to get container status \"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe\": rpc error: code = NotFound desc = could not find container \"7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe\": container with ID starting with 7c62b817fd8389d4965cf1f6d72c5cae5298b9781b45f028f82a5109cb781fbe not found: ID does not exist" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.064250 4983 scope.go:117] "RemoveContainer" containerID="a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090" Oct 01 09:33:29 crc kubenswrapper[4983]: E1001 09:33:29.064690 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090\": container with ID starting with a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090 not found: ID does not exist" containerID="a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.064728 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090"} err="failed to get container status \"a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090\": rpc error: code = NotFound desc = could not find container \"a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090\": container with ID starting with a5fbffe3101f3c2f4daa8b488dfeda9dccdd85659839e2ece270270428206090 not found: ID does not exist" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.064752 4983 scope.go:117] "RemoveContainer" containerID="192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2" Oct 01 09:33:29 crc kubenswrapper[4983]: E1001 09:33:29.065225 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2\": container with ID starting with 192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2 not found: ID does not exist" containerID="192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2" Oct 01 09:33:29 crc kubenswrapper[4983]: I1001 09:33:29.065306 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2"} err="failed to get container status \"192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2\": rpc error: code = NotFound desc = could not find container \"192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2\": container with ID starting with 192b5bc7df452925a5d63d7e5ba0890d6428d45526c9daa106329427b7a29bf2 not found: ID does not exist" Oct 01 09:33:30 crc kubenswrapper[4983]: I1001 09:33:30.132580 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:30 crc kubenswrapper[4983]: I1001 09:33:30.133004 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:30 crc kubenswrapper[4983]: I1001 09:33:30.200610 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:30 crc kubenswrapper[4983]: I1001 09:33:30.728759 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" path="/var/lib/kubelet/pods/20ab519c-9923-4ca8-a298-b0c0eea34290/volumes" Oct 01 09:33:31 crc kubenswrapper[4983]: I1001 09:33:31.053554 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:31 crc kubenswrapper[4983]: I1001 09:33:31.961055 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:33:31 crc kubenswrapper[4983]: I1001 09:33:31.961464 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:33:32 crc kubenswrapper[4983]: I1001 09:33:32.203603 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:33 crc kubenswrapper[4983]: I1001 09:33:33.016627 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zgv6c" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="registry-server" containerID="cri-o://5da91bc6d9dbf6178596c28a546c5eaa0265ba65dba6ea713a1ad6ce90a7eecd" gracePeriod=2 Oct 01 09:33:33 crc kubenswrapper[4983]: I1001 09:33:33.420063 4983 scope.go:117] "RemoveContainer" containerID="f5cf9c616c03b16b472300f1068b9e63e581b7a4115b8924061fd1137a9a77ae" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.026870 4983 generic.go:334] "Generic (PLEG): container finished" podID="9009fea7-66ed-46b2-95a7-db41ace04980" containerID="5da91bc6d9dbf6178596c28a546c5eaa0265ba65dba6ea713a1ad6ce90a7eecd" exitCode=0 Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.026922 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerDied","Data":"5da91bc6d9dbf6178596c28a546c5eaa0265ba65dba6ea713a1ad6ce90a7eecd"} Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.223968 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.334075 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content\") pod \"9009fea7-66ed-46b2-95a7-db41ace04980\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.334445 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flcqj\" (UniqueName: \"kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj\") pod \"9009fea7-66ed-46b2-95a7-db41ace04980\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.334474 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities\") pod \"9009fea7-66ed-46b2-95a7-db41ace04980\" (UID: \"9009fea7-66ed-46b2-95a7-db41ace04980\") " Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.335679 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities" (OuterVolumeSpecName: "utilities") pod "9009fea7-66ed-46b2-95a7-db41ace04980" (UID: "9009fea7-66ed-46b2-95a7-db41ace04980"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.339955 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj" (OuterVolumeSpecName: "kube-api-access-flcqj") pod "9009fea7-66ed-46b2-95a7-db41ace04980" (UID: "9009fea7-66ed-46b2-95a7-db41ace04980"). InnerVolumeSpecName "kube-api-access-flcqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.410240 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9009fea7-66ed-46b2-95a7-db41ace04980" (UID: "9009fea7-66ed-46b2-95a7-db41ace04980"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.439239 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.439321 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flcqj\" (UniqueName: \"kubernetes.io/projected/9009fea7-66ed-46b2-95a7-db41ace04980-kube-api-access-flcqj\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:34 crc kubenswrapper[4983]: I1001 09:33:34.439352 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9009fea7-66ed-46b2-95a7-db41ace04980-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.040147 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgv6c" event={"ID":"9009fea7-66ed-46b2-95a7-db41ace04980","Type":"ContainerDied","Data":"7a9ff0d166e387244a50c6374241b287cb0ba486d740a24b2083f61d934264e6"} Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.040210 4983 scope.go:117] "RemoveContainer" containerID="5da91bc6d9dbf6178596c28a546c5eaa0265ba65dba6ea713a1ad6ce90a7eecd" Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.040319 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgv6c" Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.059856 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.059981 4983 scope.go:117] "RemoveContainer" containerID="cd6b9067b6295d6e860063c71671a4a08b7448e7c6cc1bc0f7d52ec449ebe109" Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.064547 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zgv6c"] Oct 01 09:33:35 crc kubenswrapper[4983]: I1001 09:33:35.076100 4983 scope.go:117] "RemoveContainer" containerID="ac99ce21090fe78b424c3af0dbf1e909937c1627d0eb83e10c77a2371a6f122f" Oct 01 09:33:36 crc kubenswrapper[4983]: I1001 09:33:36.722704 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" path="/var/lib/kubelet/pods/9009fea7-66ed-46b2-95a7-db41ace04980/volumes" Oct 01 09:34:01 crc kubenswrapper[4983]: I1001 09:34:01.960833 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:34:01 crc kubenswrapper[4983]: I1001 09:34:01.961348 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:34:01 crc kubenswrapper[4983]: I1001 09:34:01.961389 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:34:01 crc kubenswrapper[4983]: I1001 09:34:01.962078 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:34:01 crc kubenswrapper[4983]: I1001 09:34:01.962133 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7" gracePeriod=600 Oct 01 09:34:02 crc kubenswrapper[4983]: I1001 09:34:02.253697 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7" exitCode=0 Oct 01 09:34:02 crc kubenswrapper[4983]: I1001 09:34:02.253967 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7"} Oct 01 09:34:02 crc kubenswrapper[4983]: I1001 09:34:02.254276 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerStarted","Data":"2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2"} Oct 01 09:34:02 crc kubenswrapper[4983]: I1001 09:34:02.254297 4983 scope.go:117] "RemoveContainer" containerID="d1b77b49ed7fe752b08c116dfb8cdc57a15d12fd9ec3f666a4fb4aac3093cf50" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.962592 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.963756 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.963783 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.963868 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.963884 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.963915 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="extract-utilities" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.963927 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="extract-utilities" Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.963946 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="extract-utilities" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.963959 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="extract-utilities" Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.963978 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="extract-content" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.963991 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="extract-content" Oct 01 09:34:06 crc kubenswrapper[4983]: E1001 09:34:06.964004 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="extract-content" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.964016 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="extract-content" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.964276 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="9009fea7-66ed-46b2-95a7-db41ace04980" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.964293 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ab519c-9923-4ca8-a298-b0c0eea34290" containerName="registry-server" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.966592 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:06 crc kubenswrapper[4983]: I1001 09:34:06.972996 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.031916 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.032107 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.032168 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h66hs\" (UniqueName: \"kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.133856 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h66hs\" (UniqueName: \"kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.133942 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.133999 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.134438 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.134489 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.153989 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h66hs\" (UniqueName: \"kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs\") pod \"certified-operators-bn226\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.291823 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:07 crc kubenswrapper[4983]: I1001 09:34:07.782800 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:08 crc kubenswrapper[4983]: I1001 09:34:08.298669 4983 generic.go:334] "Generic (PLEG): container finished" podID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerID="a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead" exitCode=0 Oct 01 09:34:08 crc kubenswrapper[4983]: I1001 09:34:08.298720 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerDied","Data":"a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead"} Oct 01 09:34:08 crc kubenswrapper[4983]: I1001 09:34:08.299127 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerStarted","Data":"e5df6d11fb09b5d484caeefc02fa27e3f2ba538f9f07e0be086f394d239cb654"} Oct 01 09:34:09 crc kubenswrapper[4983]: I1001 09:34:09.309627 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerStarted","Data":"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd"} Oct 01 09:34:10 crc kubenswrapper[4983]: I1001 09:34:10.323967 4983 generic.go:334] "Generic (PLEG): container finished" podID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerID="a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd" exitCode=0 Oct 01 09:34:10 crc kubenswrapper[4983]: I1001 09:34:10.324089 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerDied","Data":"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd"} Oct 01 09:34:11 crc kubenswrapper[4983]: I1001 09:34:11.332652 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerStarted","Data":"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070"} Oct 01 09:34:11 crc kubenswrapper[4983]: I1001 09:34:11.350199 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bn226" podStartSLOduration=2.702532352 podStartE2EDuration="5.350178199s" podCreationTimestamp="2025-10-01 09:34:06 +0000 UTC" firstStartedPulling="2025-10-01 09:34:08.300405018 +0000 UTC m=+1476.289633835" lastFinishedPulling="2025-10-01 09:34:10.948050875 +0000 UTC m=+1478.937279682" observedRunningTime="2025-10-01 09:34:11.345893517 +0000 UTC m=+1479.335122314" watchObservedRunningTime="2025-10-01 09:34:11.350178199 +0000 UTC m=+1479.339406996" Oct 01 09:34:17 crc kubenswrapper[4983]: I1001 09:34:17.292920 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:17 crc kubenswrapper[4983]: I1001 09:34:17.293440 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:17 crc kubenswrapper[4983]: I1001 09:34:17.368874 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:17 crc kubenswrapper[4983]: I1001 09:34:17.440537 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:17 crc kubenswrapper[4983]: I1001 09:34:17.610733 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.399690 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bn226" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="registry-server" containerID="cri-o://d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070" gracePeriod=2 Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.800304 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.941860 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities\") pod \"3bd3750c-2d78-4480-9ff5-684bd5b26419\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.942037 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content\") pod \"3bd3750c-2d78-4480-9ff5-684bd5b26419\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.942104 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h66hs\" (UniqueName: \"kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs\") pod \"3bd3750c-2d78-4480-9ff5-684bd5b26419\" (UID: \"3bd3750c-2d78-4480-9ff5-684bd5b26419\") " Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.943398 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities" (OuterVolumeSpecName: "utilities") pod "3bd3750c-2d78-4480-9ff5-684bd5b26419" (UID: "3bd3750c-2d78-4480-9ff5-684bd5b26419"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:34:19 crc kubenswrapper[4983]: I1001 09:34:19.947489 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs" (OuterVolumeSpecName: "kube-api-access-h66hs") pod "3bd3750c-2d78-4480-9ff5-684bd5b26419" (UID: "3bd3750c-2d78-4480-9ff5-684bd5b26419"). InnerVolumeSpecName "kube-api-access-h66hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.044459 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h66hs\" (UniqueName: \"kubernetes.io/projected/3bd3750c-2d78-4480-9ff5-684bd5b26419-kube-api-access-h66hs\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.044522 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.108966 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bd3750c-2d78-4480-9ff5-684bd5b26419" (UID: "3bd3750c-2d78-4480-9ff5-684bd5b26419"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.146031 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd3750c-2d78-4480-9ff5-684bd5b26419-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.413408 4983 generic.go:334] "Generic (PLEG): container finished" podID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerID="d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070" exitCode=0 Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.413457 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerDied","Data":"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070"} Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.413490 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn226" event={"ID":"3bd3750c-2d78-4480-9ff5-684bd5b26419","Type":"ContainerDied","Data":"e5df6d11fb09b5d484caeefc02fa27e3f2ba538f9f07e0be086f394d239cb654"} Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.413490 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn226" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.413513 4983 scope.go:117] "RemoveContainer" containerID="d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.450973 4983 scope.go:117] "RemoveContainer" containerID="a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.471734 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.478138 4983 scope.go:117] "RemoveContainer" containerID="a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.478424 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bn226"] Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.520954 4983 scope.go:117] "RemoveContainer" containerID="d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070" Oct 01 09:34:20 crc kubenswrapper[4983]: E1001 09:34:20.521334 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070\": container with ID starting with d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070 not found: ID does not exist" containerID="d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.521368 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070"} err="failed to get container status \"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070\": rpc error: code = NotFound desc = could not find container \"d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070\": container with ID starting with d4e82cffbeccfe32438bd6c0cc66876b53e4843ae50b63d6788c33c1c3253070 not found: ID does not exist" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.521389 4983 scope.go:117] "RemoveContainer" containerID="a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd" Oct 01 09:34:20 crc kubenswrapper[4983]: E1001 09:34:20.521777 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd\": container with ID starting with a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd not found: ID does not exist" containerID="a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.521796 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd"} err="failed to get container status \"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd\": rpc error: code = NotFound desc = could not find container \"a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd\": container with ID starting with a20ff18abb043381418cfe33f8e7c93490b8e93014766d9085137946442f0acd not found: ID does not exist" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.521842 4983 scope.go:117] "RemoveContainer" containerID="a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead" Oct 01 09:34:20 crc kubenswrapper[4983]: E1001 09:34:20.522335 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead\": container with ID starting with a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead not found: ID does not exist" containerID="a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.522356 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead"} err="failed to get container status \"a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead\": rpc error: code = NotFound desc = could not find container \"a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead\": container with ID starting with a217bc5a5f15492c4accbe07d7165b556b34135e1c754f6e0d5896ec77204ead not found: ID does not exist" Oct 01 09:34:20 crc kubenswrapper[4983]: I1001 09:34:20.724442 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" path="/var/lib/kubelet/pods/3bd3750c-2d78-4480-9ff5-684bd5b26419/volumes" Oct 01 09:34:22 crc kubenswrapper[4983]: I1001 09:34:22.051262 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7tsvg"] Oct 01 09:34:22 crc kubenswrapper[4983]: I1001 09:34:22.055939 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7tsvg"] Oct 01 09:34:22 crc kubenswrapper[4983]: I1001 09:34:22.728166 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2667ff-ed9c-4990-866d-737077afbb35" path="/var/lib/kubelet/pods/8d2667ff-ed9c-4990-866d-737077afbb35/volumes" Oct 01 09:34:32 crc kubenswrapper[4983]: I1001 09:34:32.025466 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-0cf5-account-create-gwzxr"] Oct 01 09:34:32 crc kubenswrapper[4983]: I1001 09:34:32.030720 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-0cf5-account-create-gwzxr"] Oct 01 09:34:32 crc kubenswrapper[4983]: I1001 09:34:32.737606 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139a1225-84e0-4a5c-9e2f-80f7204cde88" path="/var/lib/kubelet/pods/139a1225-84e0-4a5c-9e2f-80f7204cde88/volumes" Oct 01 09:34:33 crc kubenswrapper[4983]: I1001 09:34:33.807460 4983 scope.go:117] "RemoveContainer" containerID="ca96ad9198679e8229510fd6a04d3c26231d6483d63facb7d237d74927488021" Oct 01 09:34:33 crc kubenswrapper[4983]: I1001 09:34:33.849373 4983 scope.go:117] "RemoveContainer" containerID="0c2677dbf7d33bf4041ffdd1108f0a23339942521d75c5ee1108b528e0adb026" Oct 01 09:34:33 crc kubenswrapper[4983]: I1001 09:34:33.871315 4983 scope.go:117] "RemoveContainer" containerID="c20f42e2eae620159a19d0e70c13310bd9f15f46a9e185bb78f4e39ebbfa72b2" Oct 01 09:34:33 crc kubenswrapper[4983]: I1001 09:34:33.898643 4983 scope.go:117] "RemoveContainer" containerID="470a69ffccc2a250c4241d1a7ae1c412843d0be71c005894747d77746f4ac147" Oct 01 09:34:33 crc kubenswrapper[4983]: I1001 09:34:33.925323 4983 scope.go:117] "RemoveContainer" containerID="7cfd1c70f4e00a448514c47f43c1ea095b6f4f35d522652294041ed21f74214a" Oct 01 09:34:49 crc kubenswrapper[4983]: I1001 09:34:49.029389 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-pqcnq"] Oct 01 09:34:49 crc kubenswrapper[4983]: I1001 09:34:49.043739 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-pqcnq"] Oct 01 09:34:50 crc kubenswrapper[4983]: I1001 09:34:50.723777 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8174c1ce-cbb1-4288-a578-0478d23d3004" path="/var/lib/kubelet/pods/8174c1ce-cbb1-4288-a578-0478d23d3004/volumes" Oct 01 09:34:55 crc kubenswrapper[4983]: I1001 09:34:55.039388 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-gppdt"] Oct 01 09:34:55 crc kubenswrapper[4983]: I1001 09:34:55.045900 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-gppdt"] Oct 01 09:34:56 crc kubenswrapper[4983]: I1001 09:34:56.721540 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9bdb0d-abc4-4968-9906-8a25492d3369" path="/var/lib/kubelet/pods/ad9bdb0d-abc4-4968-9906-8a25492d3369/volumes" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.499534 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:35:23 crc kubenswrapper[4983]: E1001 09:35:23.500390 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="extract-utilities" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.500406 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="extract-utilities" Oct 01 09:35:23 crc kubenswrapper[4983]: E1001 09:35:23.500424 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="extract-content" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.500432 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="extract-content" Oct 01 09:35:23 crc kubenswrapper[4983]: E1001 09:35:23.500451 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="registry-server" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.500458 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="registry-server" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.500615 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd3750c-2d78-4480-9ff5-684bd5b26419" containerName="registry-server" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.501185 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.503377 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.504396 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-c44s5" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.504396 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.505247 4983 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.515011 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.539122 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config-secret\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.539175 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.539200 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sghlk\" (UniqueName: \"kubernetes.io/projected/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-kube-api-access-sghlk\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.539229 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-scripts\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.640778 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config-secret\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.640858 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.640894 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sghlk\" (UniqueName: \"kubernetes.io/projected/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-kube-api-access-sghlk\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.640921 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-scripts\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.642048 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-scripts\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.642051 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.647916 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-openstack-config-secret\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.657218 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sghlk\" (UniqueName: \"kubernetes.io/projected/cfb63e1a-6ced-4661-9fef-4057eb3b4e36-kube-api-access-sghlk\") pod \"openstackclient\" (UID: \"cfb63e1a-6ced-4661-9fef-4057eb3b4e36\") " pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:23 crc kubenswrapper[4983]: I1001 09:35:23.824526 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 01 09:35:24 crc kubenswrapper[4983]: I1001 09:35:24.223869 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 01 09:35:24 crc kubenswrapper[4983]: I1001 09:35:24.922670 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"cfb63e1a-6ced-4661-9fef-4057eb3b4e36","Type":"ContainerStarted","Data":"35a770bc9787a5dfb5c11d4586e500ce77381b6f84b6abe696f7b60ad3a9e3ef"} Oct 01 09:35:24 crc kubenswrapper[4983]: I1001 09:35:24.923008 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"cfb63e1a-6ced-4661-9fef-4057eb3b4e36","Type":"ContainerStarted","Data":"752beb5fdc967fcee929b1ccf777ff86edb2734ecf158dad95c07ba1e4113f3a"} Oct 01 09:35:24 crc kubenswrapper[4983]: I1001 09:35:24.935698 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.935678864 podStartE2EDuration="1.935678864s" podCreationTimestamp="2025-10-01 09:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:35:24.93439734 +0000 UTC m=+1552.923626157" watchObservedRunningTime="2025-10-01 09:35:24.935678864 +0000 UTC m=+1552.924907661" Oct 01 09:35:34 crc kubenswrapper[4983]: I1001 09:35:34.039358 4983 scope.go:117] "RemoveContainer" containerID="c063147fd5beee2beb6ae279fea7e356e634be0ba924530764e660e8173f9afe" Oct 01 09:35:34 crc kubenswrapper[4983]: I1001 09:35:34.073863 4983 scope.go:117] "RemoveContainer" containerID="b6e98d45d0505685c635bd3b875544adafd031698dd580bd2ad57c59692ea66c" Oct 01 09:35:34 crc kubenswrapper[4983]: I1001 09:35:34.116383 4983 scope.go:117] "RemoveContainer" containerID="6040c8cf1f30a295769765f93d561aa693a90d3fe36f695993c499bcc27d75bc" Oct 01 09:36:31 crc kubenswrapper[4983]: I1001 09:36:31.961510 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:36:31 crc kubenswrapper[4983]: I1001 09:36:31.962480 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.895621 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qf4nw/must-gather-tgdxv"] Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.897696 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.899411 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qf4nw"/"default-dockercfg-n5x22" Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.900864 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qf4nw"/"openshift-service-ca.crt" Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.907259 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qf4nw"/"kube-root-ca.crt" Oct 01 09:36:38 crc kubenswrapper[4983]: I1001 09:36:38.908526 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qf4nw/must-gather-tgdxv"] Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.025298 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.025372 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsppc\" (UniqueName: \"kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.126598 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsppc\" (UniqueName: \"kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.126738 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.127276 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.145979 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsppc\" (UniqueName: \"kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc\") pod \"must-gather-tgdxv\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.221563 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:36:39 crc kubenswrapper[4983]: I1001 09:36:39.684948 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qf4nw/must-gather-tgdxv"] Oct 01 09:36:40 crc kubenswrapper[4983]: I1001 09:36:40.532946 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" event={"ID":"48085d3b-ed85-4129-a15a-735beeda7f93","Type":"ContainerStarted","Data":"8c849e7a249ad3dfd209dca3c5d90bdafd310305426dd42fabc352b7cef52813"} Oct 01 09:36:44 crc kubenswrapper[4983]: I1001 09:36:44.564123 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" event={"ID":"48085d3b-ed85-4129-a15a-735beeda7f93","Type":"ContainerStarted","Data":"17c190b070665d210431a1e703575a92665876e501914b5d29760b95a5eaa8ba"} Oct 01 09:36:44 crc kubenswrapper[4983]: I1001 09:36:44.564759 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" event={"ID":"48085d3b-ed85-4129-a15a-735beeda7f93","Type":"ContainerStarted","Data":"6297f66b5ff9c9ef609f8a518a298b48e29279015185c404d29629dabb589e3b"} Oct 01 09:36:44 crc kubenswrapper[4983]: I1001 09:36:44.584889 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" podStartSLOduration=2.810825897 podStartE2EDuration="6.584864902s" podCreationTimestamp="2025-10-01 09:36:38 +0000 UTC" firstStartedPulling="2025-10-01 09:36:39.694257905 +0000 UTC m=+1627.683486702" lastFinishedPulling="2025-10-01 09:36:43.46829691 +0000 UTC m=+1631.457525707" observedRunningTime="2025-10-01 09:36:44.580202406 +0000 UTC m=+1632.569431213" watchObservedRunningTime="2025-10-01 09:36:44.584864902 +0000 UTC m=+1632.574093699" Oct 01 09:37:01 crc kubenswrapper[4983]: I1001 09:37:01.961355 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:37:01 crc kubenswrapper[4983]: I1001 09:37:01.962000 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.065399 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.206728 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.238462 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.250290 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.412114 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.415298 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.415313 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_07fc6c1979dc843040d04232d61e9581956720057e4ef0d8ce26c61d808wrp5_155ac6db-0abb-432d-bf0d-1da60a1fab4d/extract/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.581124 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.736354 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.779039 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.781244 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.942172 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/pull/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.944881 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/util/0.log" Oct 01 09:37:14 crc kubenswrapper[4983]: I1001 09:37:14.961743 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27bc8640bf9574cb4ef345ae08eb5466dd1c33b998127973c6eb353a1679tb2_7a8ec064-7744-4a7c-8479-04b47b73712a/extract/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.105085 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.252112 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/pull/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.295955 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.302664 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/pull/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.431662 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/pull/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.477203 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/extract/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.535017 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4aae4b7b802a106f1966de196fadd02ef3c779bdc9913c648dd816f23bzhp4s_c46ba0ed-905d-4f0e-a586-30c28bbdb43a/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.635095 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.765616 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/pull/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.785441 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.790814 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/pull/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.958078 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/extract/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.966696 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/util/0.log" Oct 01 09:37:15 crc kubenswrapper[4983]: I1001 09:37:15.987443 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d5cc28ad010f32ce06cf117049c4768a176a4cc2b1845b06212e1cf1356mpm_0da521dc-47ad-4443-b302-222c37f03e38/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.116615 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.267551 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.289363 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.301312 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.453158 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.492900 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/extract/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.495686 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5902qxhj_5649915e-2f6c-4e93-8aba-9820825b4643/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.499018 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.654476 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.671337 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.672928 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.840988 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/pull/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.842721 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/util/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.853748 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ccd7ba0e3641b58b769150b773d66c24839b2b4bd2cc3744868f2e60a3jfcmg_d2d46656-31cb-4e64-a979-7606de46d2fc/extract/0.log" Oct 01 09:37:16 crc kubenswrapper[4983]: I1001 09:37:16.872392 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/util/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.107305 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/pull/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.125490 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/util/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.136289 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/pull/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.277040 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/pull/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.295732 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/util/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.306042 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7e08b4ef4c4210da6849e65893e25f02a2f1e5ad24c8e4d88ab10670fj4dkn_756c9c8b-3d0a-4179-b980-aca2f443ed3f/extract/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.325053 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-bd6948d9-h6ndj_c32f9b5c-913b-4cea-8850-3c71e4ae1942/kube-rbac-proxy/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.477043 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-8xnbm_aac91626-89d0-42e2-bb69-41250a58f504/registry-server/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.491519 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-bd6948d9-h6ndj_c32f9b5c-913b-4cea-8850-3c71e4ae1942/manager/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.509999 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78d6c7d9cd-flwl2_cda8013d-e400-4905-9533-5d0187eb33b9/kube-rbac-proxy/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.737317 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78d6c7d9cd-flwl2_cda8013d-e400-4905-9533-5d0187eb33b9/manager/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.770633 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-nvtrg_ed198fb7-7179-4a0a-b8f8-829cfebc0cee/registry-server/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.807833 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d99444df-4zdf6_37538e82-453e-4a0a-858b-9929a1474605/kube-rbac-proxy/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.906574 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d99444df-4zdf6_37538e82-453e-4a0a-858b-9929a1474605/manager/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.951257 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-fxrbl_a927379f-96be-4580-ad23-d807db431881/registry-server/0.log" Oct 01 09:37:17 crc kubenswrapper[4983]: I1001 09:37:17.987657 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f8ccbb97-ppcnc_0f9b2567-8ee4-4aca-99f1-66971ccf2f2f/kube-rbac-proxy/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.091934 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-f8ccbb97-ppcnc_0f9b2567-8ee4-4aca-99f1-66971ccf2f2f/manager/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.131889 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-hsdt8_dc2d2684-81d6-4c72-9a28-7c553dda8c4c/registry-server/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.175207 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c894898-kq4lh_922be10f-7fd4-413d-a422-7b004b1be5a5/kube-rbac-proxy/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.257553 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c894898-kq4lh_922be10f-7fd4-413d-a422-7b004b1be5a5/manager/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.314353 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-2m9j5_a695e1aa-e7cd-40fd-bacb-912fc3881f9c/registry-server/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.361456 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-6k6c9_a92f120a-ff8e-479d-92ee-4ea6fc111dd0/operator/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.447214 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-xhlpw_3465433a-f044-4239-b1d6-bab1711de522/registry-server/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.525171 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7bfbc68795-lvtfh_52bd61ec-84db-4d16-8b9f-195db2a82c9c/kube-rbac-proxy/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.534370 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7bfbc68795-lvtfh_52bd61ec-84db-4d16-8b9f-195db2a82c9c/manager/0.log" Oct 01 09:37:18 crc kubenswrapper[4983]: I1001 09:37:18.634914 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-sttzg_46fb7174-5de7-47bc-829c-ef6f3c0e1ce6/registry-server/0.log" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.474765 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5svzw_05d0252d-cab0-454c-aa99-fe8f77e658bc/control-plane-machine-set-operator/0.log" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.626654 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w5pt5_65fcd0cc-38b3-4951-8379-261a96372b6f/kube-rbac-proxy/0.log" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.648397 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w5pt5_65fcd0cc-38b3-4951-8379-261a96372b6f/machine-api-operator/0.log" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.960687 4983 patch_prober.go:28] interesting pod/machine-config-daemon-pd8zz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.960740 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.960780 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.961372 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2"} pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:37:31 crc kubenswrapper[4983]: I1001 09:37:31.961417 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" containerName="machine-config-daemon" containerID="cri-o://2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" gracePeriod=600 Oct 01 09:37:32 crc kubenswrapper[4983]: E1001 09:37:32.627612 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:37:32 crc kubenswrapper[4983]: I1001 09:37:32.878351 4983 generic.go:334] "Generic (PLEG): container finished" podID="d4affe98-5451-464f-af7e-6a43e5841e02" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" exitCode=0 Oct 01 09:37:32 crc kubenswrapper[4983]: I1001 09:37:32.878404 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" event={"ID":"d4affe98-5451-464f-af7e-6a43e5841e02","Type":"ContainerDied","Data":"2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2"} Oct 01 09:37:32 crc kubenswrapper[4983]: I1001 09:37:32.878449 4983 scope.go:117] "RemoveContainer" containerID="97c94f378c99f1370d507a3a0f4f0ed6d952abd8da51c2381cf367398ff52cb7" Oct 01 09:37:32 crc kubenswrapper[4983]: I1001 09:37:32.879123 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:37:32 crc kubenswrapper[4983]: E1001 09:37:32.879542 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:37:45 crc kubenswrapper[4983]: I1001 09:37:45.621305 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-g55d4_b3825301-65ff-477d-b186-518fd24169a4/kube-rbac-proxy/0.log" Oct 01 09:37:45 crc kubenswrapper[4983]: I1001 09:37:45.660392 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-g55d4_b3825301-65ff-477d-b186-518fd24169a4/controller/0.log" Oct 01 09:37:45 crc kubenswrapper[4983]: I1001 09:37:45.804315 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-frr-files/0.log" Oct 01 09:37:45 crc kubenswrapper[4983]: I1001 09:37:45.961278 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-reloader/0.log" Oct 01 09:37:45 crc kubenswrapper[4983]: I1001 09:37:45.979559 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-frr-files/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.002984 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-reloader/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.017425 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-metrics/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.158844 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-frr-files/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.164844 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-reloader/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.183659 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-metrics/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.207145 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-metrics/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.374886 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-frr-files/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.376946 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-reloader/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.394641 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/cp-metrics/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.406417 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/controller/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.544961 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/frr-metrics/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.562961 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/kube-rbac-proxy/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.648472 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/kube-rbac-proxy-frr/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.771654 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/reloader/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.835211 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-d7w8t_97d2b4b2-bfd1-4007-949f-9f8fa01053a7/frr-k8s-webhook-server/0.log" Oct 01 09:37:46 crc kubenswrapper[4983]: I1001 09:37:46.932366 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-q6g4h_b5648467-cebb-4305-a03b-cfe484b92d9e/frr/0.log" Oct 01 09:37:47 crc kubenswrapper[4983]: I1001 09:37:47.000415 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d59c47655-9kwhv_31e9843f-42b4-4229-bdfe-d64ccd1b37eb/manager/0.log" Oct 01 09:37:47 crc kubenswrapper[4983]: I1001 09:37:47.125387 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-67c678544c-thtq2_7a7230d0-bbf6-4949-9af0-6489822f46f5/webhook-server/0.log" Oct 01 09:37:47 crc kubenswrapper[4983]: I1001 09:37:47.169077 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7cfp_13e9026f-afa7-4216-ae81-208fb18c8d08/kube-rbac-proxy/0.log" Oct 01 09:37:47 crc kubenswrapper[4983]: I1001 09:37:47.341259 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h7cfp_13e9026f-afa7-4216-ae81-208fb18c8d08/speaker/0.log" Oct 01 09:37:47 crc kubenswrapper[4983]: I1001 09:37:47.714807 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:37:47 crc kubenswrapper[4983]: E1001 09:37:47.715059 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:37:59 crc kubenswrapper[4983]: I1001 09:37:59.573531 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-c322-account-create-ttpj4_d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339/mariadb-account-create/0.log" Oct 01 09:37:59 crc kubenswrapper[4983]: I1001 09:37:59.707195 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-external-api-0-cleaner-2932185rgpqb_eebb3506-fc09-4c67-bd25-df8ee0cc59ad/glance-cache-glance-default-external-api-0-cleaner/0.log" Oct 01 09:37:59 crc kubenswrapper[4983]: I1001 09:37:59.781769 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-internal-api-0-cleaner-2932185kt5w8_9c9195c3-eb97-4802-9026-f496fd13e680/glance-cache-glance-default-internal-api-0-cleaner/0.log" Oct 01 09:37:59 crc kubenswrapper[4983]: I1001 09:37:59.877192 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-zgwdz_f0301a58-c508-4ba4-8c7a-d0e3f845bba9/mariadb-database-create/0.log" Oct 01 09:37:59 crc kubenswrapper[4983]: I1001 09:37:59.973662 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-5khjf_f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf/glance-db-sync/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.072825 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_28b127a4-30d1-44a6-ac91-d3bb63546626/glance-api/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.112515 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_28b127a4-30d1-44a6-ac91-d3bb63546626/glance-httpd/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.135477 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_28b127a4-30d1-44a6-ac91-d3bb63546626/glance-log/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.265966 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3dc1ff50-c96e-4360-a2c4-954f495f107a/glance-api/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.309856 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3dc1ff50-c96e-4360-a2c4-954f495f107a/glance-httpd/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.311557 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_3dc1ff50-c96e-4360-a2c4-954f495f107a/glance-log/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.608733 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_d44fa8ad-8a99-41d6-8dba-11911de2c065/memcached/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.750333 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-74bb96697f-mfssq_76cb22c5-718c-4303-80ce-2e15df680917/keystone-api/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.769363 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_ce45f506-705c-4232-a029-bd829ad6cae9/mysql-bootstrap/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.950742 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_ce45f506-705c-4232-a029-bd829ad6cae9/galera/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.963263 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_ce45f506-705c-4232-a029-bd829ad6cae9/mysql-bootstrap/0.log" Oct 01 09:38:00 crc kubenswrapper[4983]: I1001 09:38:00.968756 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_f6110242-9584-4645-909c-514c48956b2d/mysql-bootstrap/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.184069 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_f6110242-9584-4645-909c-514c48956b2d/mysql-bootstrap/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.201860 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_f6110242-9584-4645-909c-514c48956b2d/galera/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.215158 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_0cd75293-88f6-41e1-8d95-5b89e5a40f03/mysql-bootstrap/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.438096 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_0cd75293-88f6-41e1-8d95-5b89e5a40f03/galera/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.450577 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_cfb63e1a-6ced-4661-9fef-4057eb3b4e36/openstackclient/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.457444 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_0cd75293-88f6-41e1-8d95-5b89e5a40f03/mysql-bootstrap/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.600976 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bb4c18ce-6256-49d4-8917-e50a2a87180b/setup-container/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.772635 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bb4c18ce-6256-49d4-8917-e50a2a87180b/rabbitmq/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.811229 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bb4c18ce-6256-49d4-8917-e50a2a87180b/setup-container/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.833891 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-7578798499-ms7kc_c7f11862-e033-4205-90d6-44096017084e/proxy-httpd/0.log" Oct 01 09:38:01 crc kubenswrapper[4983]: I1001 09:38:01.963080 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-7578798499-ms7kc_c7f11862-e033-4205-90d6-44096017084e/proxy-server/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.015338 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-8thz2_3e0c9600-5b4d-4b2d-98c4-300502282d3a/swift-ring-rebalance/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.156283 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/account-auditor/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.169146 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/account-reaper/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.194089 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/account-replicator/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.211487 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/account-server/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.361393 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/container-replicator/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.377603 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/container-auditor/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.406517 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/container-server/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.408635 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/container-updater/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.660515 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/object-expirer/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.687169 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/object-auditor/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.707264 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/object-replicator/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.722513 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:38:02 crc kubenswrapper[4983]: E1001 09:38:02.722727 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.758114 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/object-updater/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.758246 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/object-server/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.815522 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/rsync/0.log" Oct 01 09:38:02 crc kubenswrapper[4983]: I1001 09:38:02.841437 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_e1456183-06e2-4b5c-9ab8-c7f0986ccec6/swift-recon-cron/0.log" Oct 01 09:38:13 crc kubenswrapper[4983]: I1001 09:38:13.709184 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-utilities/0.log" Oct 01 09:38:13 crc kubenswrapper[4983]: I1001 09:38:13.909470 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-utilities/0.log" Oct 01 09:38:13 crc kubenswrapper[4983]: I1001 09:38:13.924873 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-content/0.log" Oct 01 09:38:13 crc kubenswrapper[4983]: I1001 09:38:13.932679 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-content/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.085781 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-content/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.090428 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/extract-utilities/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.313281 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-utilities/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.406029 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4f8xr_b4a11c9b-0ffa-4c26-bf11-f3b53e1fca63/registry-server/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.495280 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-utilities/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.556611 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-content/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.575678 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-content/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.714803 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:38:14 crc kubenswrapper[4983]: E1001 09:38:14.715044 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.732199 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-content/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.733351 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/extract-utilities/0.log" Oct 01 09:38:14 crc kubenswrapper[4983]: I1001 09:38:14.994221 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/util/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.090065 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/pull/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.146651 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/util/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.159459 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cg4hg_c982d336-6845-4efb-8a7b-202f55f55e8a/registry-server/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.216339 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/pull/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.336856 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/pull/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.369676 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/util/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.377246 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d962tdf6_a7fe723e-4359-4361-ba69-cb6411ffbbf3/extract/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.503957 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ddvpk_5d2dbdcc-216c-4e21-8da0-3291670f7555/marketplace-operator/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.568157 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-utilities/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.742890 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-utilities/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.787112 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-content/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.796750 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-content/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.887654 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-utilities/0.log" Oct 01 09:38:15 crc kubenswrapper[4983]: I1001 09:38:15.912228 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/extract-content/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.048049 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8lq6m_62d5d2de-567c-4077-8d82-32833bfb6129/registry-server/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.073719 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-utilities/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.226516 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-content/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.241612 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-content/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.250278 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-utilities/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.417659 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-utilities/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.424037 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/extract-content/0.log" Oct 01 09:38:16 crc kubenswrapper[4983]: I1001 09:38:16.804573 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fcr4m_0ccb3b8a-516c-4c88-8350-21a238b80504/registry-server/0.log" Oct 01 09:38:25 crc kubenswrapper[4983]: I1001 09:38:25.715414 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:38:25 crc kubenswrapper[4983]: E1001 09:38:25.716195 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:38:36 crc kubenswrapper[4983]: I1001 09:38:36.714509 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:38:36 crc kubenswrapper[4983]: E1001 09:38:36.716490 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:38:44 crc kubenswrapper[4983]: I1001 09:38:44.053404 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-zgwdz"] Oct 01 09:38:44 crc kubenswrapper[4983]: I1001 09:38:44.061369 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-zgwdz"] Oct 01 09:38:44 crc kubenswrapper[4983]: I1001 09:38:44.722138 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0301a58-c508-4ba4-8c7a-d0e3f845bba9" path="/var/lib/kubelet/pods/f0301a58-c508-4ba4-8c7a-d0e3f845bba9/volumes" Oct 01 09:38:50 crc kubenswrapper[4983]: I1001 09:38:50.715136 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:38:50 crc kubenswrapper[4983]: E1001 09:38:50.715611 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:38:54 crc kubenswrapper[4983]: I1001 09:38:54.032908 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-c322-account-create-ttpj4"] Oct 01 09:38:54 crc kubenswrapper[4983]: I1001 09:38:54.037283 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-c322-account-create-ttpj4"] Oct 01 09:38:54 crc kubenswrapper[4983]: I1001 09:38:54.723287 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339" path="/var/lib/kubelet/pods/d7d8a3b9-3e6c-43fd-8f73-31bcb52d4339/volumes" Oct 01 09:39:01 crc kubenswrapper[4983]: I1001 09:39:01.714902 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:39:01 crc kubenswrapper[4983]: E1001 09:39:01.715736 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:39:02 crc kubenswrapper[4983]: I1001 09:39:02.033257 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5khjf"] Oct 01 09:39:02 crc kubenswrapper[4983]: I1001 09:39:02.045371 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-5khjf"] Oct 01 09:39:02 crc kubenswrapper[4983]: I1001 09:39:02.733304 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf" path="/var/lib/kubelet/pods/f774b9b8-52f2-4c3b-a6f7-16aafc4c3bdf/volumes" Oct 01 09:39:12 crc kubenswrapper[4983]: I1001 09:39:12.724923 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:39:12 crc kubenswrapper[4983]: E1001 09:39:12.725878 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:39:15 crc kubenswrapper[4983]: I1001 09:39:15.684776 4983 generic.go:334] "Generic (PLEG): container finished" podID="48085d3b-ed85-4129-a15a-735beeda7f93" containerID="6297f66b5ff9c9ef609f8a518a298b48e29279015185c404d29629dabb589e3b" exitCode=0 Oct 01 09:39:15 crc kubenswrapper[4983]: I1001 09:39:15.684864 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" event={"ID":"48085d3b-ed85-4129-a15a-735beeda7f93","Type":"ContainerDied","Data":"6297f66b5ff9c9ef609f8a518a298b48e29279015185c404d29629dabb589e3b"} Oct 01 09:39:15 crc kubenswrapper[4983]: I1001 09:39:15.685548 4983 scope.go:117] "RemoveContainer" containerID="6297f66b5ff9c9ef609f8a518a298b48e29279015185c404d29629dabb589e3b" Oct 01 09:39:16 crc kubenswrapper[4983]: I1001 09:39:16.637426 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qf4nw_must-gather-tgdxv_48085d3b-ed85-4129-a15a-735beeda7f93/gather/0.log" Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.441689 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qf4nw/must-gather-tgdxv"] Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.442426 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" podUID="48085d3b-ed85-4129-a15a-735beeda7f93" containerName="copy" containerID="cri-o://17c190b070665d210431a1e703575a92665876e501914b5d29760b95a5eaa8ba" gracePeriod=2 Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.446859 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qf4nw/must-gather-tgdxv"] Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.755142 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qf4nw_must-gather-tgdxv_48085d3b-ed85-4129-a15a-735beeda7f93/copy/0.log" Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.760866 4983 generic.go:334] "Generic (PLEG): container finished" podID="48085d3b-ed85-4129-a15a-735beeda7f93" containerID="17c190b070665d210431a1e703575a92665876e501914b5d29760b95a5eaa8ba" exitCode=143 Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.761025 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c849e7a249ad3dfd209dca3c5d90bdafd310305426dd42fabc352b7cef52813" Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.823488 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qf4nw_must-gather-tgdxv_48085d3b-ed85-4129-a15a-735beeda7f93/copy/0.log" Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.824123 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.963745 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output\") pod \"48085d3b-ed85-4129-a15a-735beeda7f93\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.964045 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsppc\" (UniqueName: \"kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc\") pod \"48085d3b-ed85-4129-a15a-735beeda7f93\" (UID: \"48085d3b-ed85-4129-a15a-735beeda7f93\") " Oct 01 09:39:23 crc kubenswrapper[4983]: I1001 09:39:23.970539 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc" (OuterVolumeSpecName: "kube-api-access-fsppc") pod "48085d3b-ed85-4129-a15a-735beeda7f93" (UID: "48085d3b-ed85-4129-a15a-735beeda7f93"). InnerVolumeSpecName "kube-api-access-fsppc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:24 crc kubenswrapper[4983]: I1001 09:39:24.033049 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "48085d3b-ed85-4129-a15a-735beeda7f93" (UID: "48085d3b-ed85-4129-a15a-735beeda7f93"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:24 crc kubenswrapper[4983]: I1001 09:39:24.066079 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsppc\" (UniqueName: \"kubernetes.io/projected/48085d3b-ed85-4129-a15a-735beeda7f93-kube-api-access-fsppc\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:24 crc kubenswrapper[4983]: I1001 09:39:24.066120 4983 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/48085d3b-ed85-4129-a15a-735beeda7f93-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:24 crc kubenswrapper[4983]: I1001 09:39:24.723072 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48085d3b-ed85-4129-a15a-735beeda7f93" path="/var/lib/kubelet/pods/48085d3b-ed85-4129-a15a-735beeda7f93/volumes" Oct 01 09:39:24 crc kubenswrapper[4983]: I1001 09:39:24.767873 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qf4nw/must-gather-tgdxv" Oct 01 09:39:27 crc kubenswrapper[4983]: I1001 09:39:27.715302 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:39:27 crc kubenswrapper[4983]: E1001 09:39:27.716347 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:39:34 crc kubenswrapper[4983]: I1001 09:39:34.233784 4983 scope.go:117] "RemoveContainer" containerID="4b755735c16e1f9f98f7038c8cd883ad33e338aa935bcd4c0f060c34e39d4ca6" Oct 01 09:39:34 crc kubenswrapper[4983]: I1001 09:39:34.254871 4983 scope.go:117] "RemoveContainer" containerID="5f7cf39ad20a89f71338bf4c2cca4900fd24a57352121b6ab5f44e8af7049ecf" Oct 01 09:39:34 crc kubenswrapper[4983]: I1001 09:39:34.287272 4983 scope.go:117] "RemoveContainer" containerID="e7c59ebf221c2973d554934f4135cdee58ba2881abba30b0ac0df420eff0f9ca" Oct 01 09:39:42 crc kubenswrapper[4983]: I1001 09:39:42.721369 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:39:42 crc kubenswrapper[4983]: E1001 09:39:42.722083 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:39:54 crc kubenswrapper[4983]: I1001 09:39:54.714574 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:39:54 crc kubenswrapper[4983]: E1001 09:39:54.715408 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" Oct 01 09:40:05 crc kubenswrapper[4983]: I1001 09:40:05.714581 4983 scope.go:117] "RemoveContainer" containerID="2dea5169d806ed5a9f951f38ef3ee3fbf3fd4799beee3c6f7f549a44d1eb5fb2" Oct 01 09:40:05 crc kubenswrapper[4983]: E1001 09:40:05.715258 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pd8zz_openshift-machine-config-operator(d4affe98-5451-464f-af7e-6a43e5841e02)\"" pod="openshift-machine-config-operator/machine-config-daemon-pd8zz" podUID="d4affe98-5451-464f-af7e-6a43e5841e02" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067173401024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067173401017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067167443016522 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067167444015473 5ustar corecore